var/home/core/zuul-output/0000755000175000017500000000000015111315631014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111322047015465 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003736406515111322040017700 0ustar rootrootNov 25 12:09:08 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 12:09:08 crc restorecon[4704]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:08 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 12:09:09 crc restorecon[4704]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 12:09:10 crc kubenswrapper[4715]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 12:09:10 crc kubenswrapper[4715]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 12:09:10 crc kubenswrapper[4715]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 12:09:10 crc kubenswrapper[4715]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 12:09:10 crc kubenswrapper[4715]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 12:09:10 crc kubenswrapper[4715]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.195211 4715 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198435 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198454 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198460 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198466 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198470 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198474 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198477 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198487 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198492 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198497 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198501 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198504 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198508 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198512 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198515 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198524 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198527 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198532 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198537 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198540 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198544 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198547 4715 feature_gate.go:330] unrecognized feature gate: Example Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198551 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198555 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198558 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198562 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198565 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198570 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198574 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198579 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198582 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198586 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198590 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198595 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198599 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198602 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198606 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198609 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198612 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198616 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198619 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198623 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198634 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198638 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198642 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198646 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198649 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198653 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198656 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198659 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198663 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198666 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198670 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198674 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198679 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198682 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198686 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198690 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198694 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198697 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198701 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198705 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198708 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198711 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198715 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198719 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198722 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198725 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198729 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198732 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.198736 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249267 4715 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249348 4715 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249405 4715 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249420 4715 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249434 4715 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249445 4715 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249458 4715 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249487 4715 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249497 4715 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249505 4715 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249515 4715 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249522 4715 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249529 4715 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249536 4715 flags.go:64] FLAG: --cgroup-root="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249542 4715 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249548 4715 flags.go:64] FLAG: --client-ca-file="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249554 4715 flags.go:64] FLAG: --cloud-config="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249566 4715 flags.go:64] FLAG: --cloud-provider="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249571 4715 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249586 4715 flags.go:64] FLAG: --cluster-domain="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249592 4715 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249601 4715 flags.go:64] FLAG: --config-dir="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249607 4715 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249613 4715 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249622 4715 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249628 4715 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249635 4715 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249642 4715 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249648 4715 flags.go:64] FLAG: --contention-profiling="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249654 4715 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249661 4715 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249667 4715 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249673 4715 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249684 4715 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249691 4715 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249697 4715 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249704 4715 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249711 4715 flags.go:64] FLAG: --enable-server="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249717 4715 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249730 4715 flags.go:64] FLAG: --event-burst="100" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249737 4715 flags.go:64] FLAG: --event-qps="50" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249743 4715 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249751 4715 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249765 4715 flags.go:64] FLAG: --eviction-hard="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249773 4715 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249780 4715 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249787 4715 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249794 4715 flags.go:64] FLAG: --eviction-soft="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249800 4715 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249806 4715 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249812 4715 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249818 4715 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249824 4715 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249830 4715 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249836 4715 flags.go:64] FLAG: --feature-gates="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249843 4715 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249849 4715 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249856 4715 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249862 4715 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249868 4715 flags.go:64] FLAG: --healthz-port="10248" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249874 4715 flags.go:64] FLAG: --help="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249880 4715 flags.go:64] FLAG: --hostname-override="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249886 4715 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249892 4715 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249898 4715 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249905 4715 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249911 4715 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249917 4715 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249923 4715 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249931 4715 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249936 4715 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249942 4715 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249949 4715 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249957 4715 flags.go:64] FLAG: --kube-reserved="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249963 4715 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249969 4715 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249975 4715 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249981 4715 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.249987 4715 flags.go:64] FLAG: --lock-file="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250002 4715 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250008 4715 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250015 4715 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250025 4715 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250031 4715 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250038 4715 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250043 4715 flags.go:64] FLAG: --logging-format="text" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250050 4715 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250057 4715 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250062 4715 flags.go:64] FLAG: --manifest-url="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250069 4715 flags.go:64] FLAG: --manifest-url-header="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250079 4715 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250085 4715 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250093 4715 flags.go:64] FLAG: --max-pods="110" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250099 4715 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250107 4715 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250113 4715 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250119 4715 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250126 4715 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250132 4715 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250139 4715 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250157 4715 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250163 4715 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250169 4715 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250176 4715 flags.go:64] FLAG: --pod-cidr="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250182 4715 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250229 4715 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250235 4715 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250241 4715 flags.go:64] FLAG: --pods-per-core="0" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250247 4715 flags.go:64] FLAG: --port="10250" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250254 4715 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250259 4715 flags.go:64] FLAG: --provider-id="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250265 4715 flags.go:64] FLAG: --qos-reserved="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250271 4715 flags.go:64] FLAG: --read-only-port="10255" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250277 4715 flags.go:64] FLAG: --register-node="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250283 4715 flags.go:64] FLAG: --register-schedulable="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250297 4715 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250308 4715 flags.go:64] FLAG: --registry-burst="10" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250314 4715 flags.go:64] FLAG: --registry-qps="5" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250320 4715 flags.go:64] FLAG: --reserved-cpus="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250327 4715 flags.go:64] FLAG: --reserved-memory="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250335 4715 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250341 4715 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250347 4715 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250353 4715 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250359 4715 flags.go:64] FLAG: --runonce="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250365 4715 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250374 4715 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250386 4715 flags.go:64] FLAG: --seccomp-default="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250400 4715 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250408 4715 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250417 4715 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250426 4715 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250434 4715 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250441 4715 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250449 4715 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250457 4715 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250464 4715 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250472 4715 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250480 4715 flags.go:64] FLAG: --system-cgroups="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250488 4715 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250500 4715 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250506 4715 flags.go:64] FLAG: --tls-cert-file="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250512 4715 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250530 4715 flags.go:64] FLAG: --tls-min-version="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250540 4715 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250555 4715 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250564 4715 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250571 4715 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250580 4715 flags.go:64] FLAG: --v="2" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250592 4715 flags.go:64] FLAG: --version="false" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250604 4715 flags.go:64] FLAG: --vmodule="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250633 4715 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.250644 4715 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257102 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257152 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257158 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257170 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257174 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257180 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257186 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257203 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257208 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257212 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257216 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257220 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257224 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257228 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257233 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257237 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257247 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257251 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257255 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257259 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257263 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257267 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257272 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257276 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257280 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257284 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257289 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257293 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257297 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257304 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257308 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257315 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257324 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257329 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257334 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257338 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257345 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257351 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257357 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257362 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257369 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257375 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257380 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257385 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257390 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257394 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257399 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257405 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257410 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257414 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257418 4715 feature_gate.go:330] unrecognized feature gate: Example Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257422 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257426 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257432 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257436 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257442 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257446 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257450 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257454 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257458 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257463 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257467 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257471 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257475 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257479 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257487 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257492 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257495 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257500 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257505 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.257512 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.257523 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.275060 4715 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.275130 4715 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275218 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275227 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275231 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275235 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275240 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275246 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275252 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275259 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275266 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275272 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275277 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275281 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275285 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275289 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275293 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275298 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275301 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275305 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275308 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275312 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275316 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275319 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275323 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275326 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275330 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275334 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275339 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275342 4715 feature_gate.go:330] unrecognized feature gate: Example Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275346 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275349 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275353 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275357 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275361 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275365 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275369 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275373 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275376 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275380 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275384 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275387 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275391 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275396 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275399 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275404 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275408 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275412 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275416 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275420 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275439 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275444 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275447 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275452 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275456 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275460 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275463 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275467 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275472 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275476 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275480 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275483 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275487 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275491 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275494 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275499 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275503 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275507 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275512 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275516 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275520 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275524 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275528 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.275535 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275661 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275667 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275672 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275676 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275681 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275685 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275689 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275693 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275697 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275701 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275705 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275710 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275713 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275717 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275721 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275725 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275728 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275732 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275735 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275739 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275743 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275746 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275749 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275753 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275757 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275761 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275765 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275768 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275772 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275777 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275782 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275786 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275789 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275793 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275797 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275802 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275806 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275809 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275813 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275816 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275820 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275825 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275829 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275833 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275838 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275841 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275847 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275851 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275856 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275861 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275866 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275871 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275875 4715 feature_gate.go:330] unrecognized feature gate: Example Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275880 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275884 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275888 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275892 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275897 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275901 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275905 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275909 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275913 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275917 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275920 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275923 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275927 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275931 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275935 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275939 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275942 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.275946 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.275952 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.276139 4715 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.281029 4715 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.281130 4715 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.283418 4715 server.go:997] "Starting client certificate rotation" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.283445 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.283650 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-14 04:35:15.88707335 +0000 UTC Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.283777 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.345561 4715 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.347466 4715 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.420928 4715 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.442447 4715 log.go:25] "Validated CRI v1 runtime API" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.562882 4715 log.go:25] "Validated CRI v1 image API" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.564408 4715 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.572354 4715 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-12-04-38-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.572388 4715 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.592348 4715 manager.go:217] Machine: {Timestamp:2025-11-25 12:09:10.587672209 +0000 UTC m=+1.095175260 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:aa0ef64c-25d2-4dc4-b8f3-98dcad372584 BootID:b3d33238-179c-4051-9790-e67bba33905c Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e3:9c:41 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e3:9c:41 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:5b:eb:aa Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:19:b3:ec Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c2:0e:2c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:f0:a0:ec Speed:-1 Mtu:1496} {Name:eth10 MacAddress:56:be:35:17:dc:df Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:7a:bc:11:9b:3e:ff Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.592595 4715 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.592747 4715 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.593933 4715 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.594174 4715 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.594246 4715 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.594520 4715 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.594533 4715 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.595241 4715 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.595858 4715 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.597103 4715 state_mem.go:36] "Initialized new in-memory state store" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.597234 4715 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.604733 4715 kubelet.go:418] "Attempting to sync node with API server" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.604767 4715 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.604825 4715 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.604846 4715 kubelet.go:324] "Adding apiserver pod source" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.604863 4715 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.610143 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.610253 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.610335 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.610449 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.610870 4715 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.612500 4715 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.621474 4715 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623781 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623810 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623818 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623827 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623840 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623857 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623868 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623884 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623895 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623904 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623919 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.623929 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.624754 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.625310 4715 server.go:1280] "Started kubelet" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.626521 4715 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.626528 4715 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.626887 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:10 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.627509 4715 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.629606 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.629680 4715 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.629849 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 18:27:14.275950172 +0000 UTC Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.629991 4715 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.630018 4715 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.630066 4715 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.630139 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.640537 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="200ms" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.640682 4715 factory.go:55] Registering systemd factory Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.640707 4715 factory.go:221] Registration of the systemd container factory successfully Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.642173 4715 factory.go:153] Registering CRI-O factory Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.642226 4715 factory.go:221] Registration of the crio container factory successfully Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.642337 4715 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.642240 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.642413 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.642363 4715 factory.go:103] Registering Raw factory Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.642506 4715 manager.go:1196] Started watching for new ooms in manager Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.642956 4715 server.go:460] "Adding debug handlers to kubelet server" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.648049 4715 manager.go:319] Starting recovery of all containers Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.648359 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.40:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b3ea36a0948d9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 12:09:10.625265881 +0000 UTC m=+1.132768902,LastTimestamp:2025-11-25 12:09:10.625265881 +0000 UTC m=+1.132768902,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.657986 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658070 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658091 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658107 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658123 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658142 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658158 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658176 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658227 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658244 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658393 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658411 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658434 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658459 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658492 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658509 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658528 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658543 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658563 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658582 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658600 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658621 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658639 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658655 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658672 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658690 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658749 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658791 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658810 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658832 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658849 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658863 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658876 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658889 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658901 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658913 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658925 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658938 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658952 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658964 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658978 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.658990 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659004 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659018 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659031 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659047 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659062 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659076 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659141 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659158 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659172 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659208 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659226 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659241 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659255 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659271 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659295 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659311 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659326 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659341 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659356 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659372 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659386 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659428 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659445 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659459 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659471 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659484 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659498 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659511 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659524 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659538 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.659552 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661813 4715 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661862 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661881 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661898 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661915 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661930 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661943 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661957 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661971 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661983 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.661998 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662011 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662026 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662040 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662054 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662067 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662080 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662093 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662105 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662117 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662136 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662149 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662160 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662173 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662207 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662221 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662233 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662246 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662257 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662283 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662297 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662310 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662333 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662348 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662363 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662379 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662394 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662410 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662423 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662436 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662449 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662462 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662475 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662488 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662499 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662510 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662523 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662536 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662547 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662559 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662573 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662585 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662605 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662617 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662628 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662642 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662653 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662667 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662679 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662690 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662704 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662716 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662727 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662740 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662751 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662765 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662813 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662829 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662842 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662855 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662868 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662880 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662893 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662906 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662919 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662931 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662945 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662959 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662972 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662985 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.662998 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663013 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663026 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663040 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663057 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663069 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663084 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663099 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663112 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663125 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663139 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663154 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663168 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663202 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663218 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663230 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663242 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663257 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663270 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663283 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663297 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663310 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663324 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663337 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663350 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663363 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663377 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663390 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663402 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663420 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663435 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663449 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663463 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663477 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663495 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663508 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663521 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663536 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663549 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663562 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663575 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663590 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663604 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663616 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663628 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663641 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663654 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663665 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663677 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663689 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663699 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663710 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663725 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663738 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663750 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663763 4715 reconstruct.go:97] "Volume reconstruction finished" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.663773 4715 reconciler.go:26] "Reconciler: start to sync state" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.671536 4715 manager.go:324] Recovery completed Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.685664 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.687576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.687657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.687672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.688841 4715 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.688859 4715 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.688887 4715 state_mem.go:36] "Initialized new in-memory state store" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.689654 4715 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.692018 4715 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.692328 4715 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.692572 4715 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.692691 4715 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 12:09:10 crc kubenswrapper[4715]: W1125 12:09:10.696413 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.696622 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.730710 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.779740 4715 policy_none.go:49] "None policy: Start" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.781379 4715 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.781425 4715 state_mem.go:35] "Initializing new in-memory state store" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.793716 4715 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.831146 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.842080 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="400ms" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.931323 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.955490 4715 manager.go:334] "Starting Device Plugin manager" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.955547 4715 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.955560 4715 server.go:79] "Starting device plugin registration server" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.956100 4715 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.956366 4715 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.956645 4715 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.956937 4715 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.956954 4715 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 12:09:10 crc kubenswrapper[4715]: E1125 12:09:10.964838 4715 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.994924 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.995135 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.996925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.996973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.996987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.997170 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.997583 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.997678 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.998096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.998129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.998141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.998288 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.998436 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.998480 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.998694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.998730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.998743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.999203 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.999237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.999271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.999389 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.999448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.999474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.999488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.999629 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 12:09:10 crc kubenswrapper[4715]: I1125 12:09:10.999693 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.000176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.000219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.000234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.000369 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.000521 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.000556 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.000752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.000786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.000809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.001132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.001156 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.001164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.001921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.001966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.001999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.002272 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.002317 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.003139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.003178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.003207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.056754 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.058096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.058137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.058150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.058175 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 12:09:11 crc kubenswrapper[4715]: E1125 12:09:11.058864 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.40:6443: connect: connection refused" node="crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069657 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069711 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069742 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069769 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069825 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069851 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069878 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069899 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069920 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069942 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069963 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.069989 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.070078 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.070149 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.070175 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171326 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171438 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171464 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171482 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171497 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171513 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171526 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171541 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171557 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171572 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171590 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171620 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171635 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.171654 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172180 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172273 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172304 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172339 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172364 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172388 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172418 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172446 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172454 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172473 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172500 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172505 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172494 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172536 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.172526 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: E1125 12:09:11.243582 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="800ms" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.259675 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.261437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.261491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.261501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.261561 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 12:09:11 crc kubenswrapper[4715]: E1125 12:09:11.262210 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.40:6443: connect: connection refused" node="crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.331093 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.339876 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.356162 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.379509 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.388824 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:11 crc kubenswrapper[4715]: W1125 12:09:11.460398 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c1d867613ab6b40c06ab80ff39a8164e40a1d092e16fe97110fdd1dc3e721563 WatchSource:0}: Error finding container c1d867613ab6b40c06ab80ff39a8164e40a1d092e16fe97110fdd1dc3e721563: Status 404 returned error can't find the container with id c1d867613ab6b40c06ab80ff39a8164e40a1d092e16fe97110fdd1dc3e721563 Nov 25 12:09:11 crc kubenswrapper[4715]: W1125 12:09:11.467031 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-3040a0865beea78799925d2b7a11e3f38a04d4a8341627461faed15d088ba552 WatchSource:0}: Error finding container 3040a0865beea78799925d2b7a11e3f38a04d4a8341627461faed15d088ba552: Status 404 returned error can't find the container with id 3040a0865beea78799925d2b7a11e3f38a04d4a8341627461faed15d088ba552 Nov 25 12:09:11 crc kubenswrapper[4715]: W1125 12:09:11.469324 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4983456ff0c182773d6e3f2f8398679fdf4c78966682f1e7f9565124870adcd9 WatchSource:0}: Error finding container 4983456ff0c182773d6e3f2f8398679fdf4c78966682f1e7f9565124870adcd9: Status 404 returned error can't find the container with id 4983456ff0c182773d6e3f2f8398679fdf4c78966682f1e7f9565124870adcd9 Nov 25 12:09:11 crc kubenswrapper[4715]: W1125 12:09:11.473533 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-3744cd2e253199d31b53b3539fd1f9e3117e9b86ef942133948db30311f1c05d WatchSource:0}: Error finding container 3744cd2e253199d31b53b3539fd1f9e3117e9b86ef942133948db30311f1c05d: Status 404 returned error can't find the container with id 3744cd2e253199d31b53b3539fd1f9e3117e9b86ef942133948db30311f1c05d Nov 25 12:09:11 crc kubenswrapper[4715]: W1125 12:09:11.483220 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-0adb16b7b96a93fbc2f7bf2d40d5f689c693ac8b0ccd932098ae9f0d48765f34 WatchSource:0}: Error finding container 0adb16b7b96a93fbc2f7bf2d40d5f689c693ac8b0ccd932098ae9f0d48765f34: Status 404 returned error can't find the container with id 0adb16b7b96a93fbc2f7bf2d40d5f689c693ac8b0ccd932098ae9f0d48765f34 Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.628427 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.630535 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 14:30:11.056284604 +0000 UTC Nov 25 12:09:11 crc kubenswrapper[4715]: W1125 12:09:11.633459 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:11 crc kubenswrapper[4715]: E1125 12:09:11.633551 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.663283 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.664918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.664987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.665016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.665051 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 12:09:11 crc kubenswrapper[4715]: E1125 12:09:11.665769 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.40:6443: connect: connection refused" node="crc" Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.697690 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3040a0865beea78799925d2b7a11e3f38a04d4a8341627461faed15d088ba552"} Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.698819 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c1d867613ab6b40c06ab80ff39a8164e40a1d092e16fe97110fdd1dc3e721563"} Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.699905 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0adb16b7b96a93fbc2f7bf2d40d5f689c693ac8b0ccd932098ae9f0d48765f34"} Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.701080 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3744cd2e253199d31b53b3539fd1f9e3117e9b86ef942133948db30311f1c05d"} Nov 25 12:09:11 crc kubenswrapper[4715]: I1125 12:09:11.702539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4983456ff0c182773d6e3f2f8398679fdf4c78966682f1e7f9565124870adcd9"} Nov 25 12:09:11 crc kubenswrapper[4715]: W1125 12:09:11.704406 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:11 crc kubenswrapper[4715]: E1125 12:09:11.704508 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:11 crc kubenswrapper[4715]: W1125 12:09:11.981056 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:11 crc kubenswrapper[4715]: E1125 12:09:11.981170 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:12 crc kubenswrapper[4715]: E1125 12:09:12.044903 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="1.6s" Nov 25 12:09:12 crc kubenswrapper[4715]: W1125 12:09:12.056020 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:12 crc kubenswrapper[4715]: E1125 12:09:12.056142 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:12 crc kubenswrapper[4715]: I1125 12:09:12.466370 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:12 crc kubenswrapper[4715]: I1125 12:09:12.467890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:12 crc kubenswrapper[4715]: I1125 12:09:12.467916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:12 crc kubenswrapper[4715]: I1125 12:09:12.467926 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:12 crc kubenswrapper[4715]: I1125 12:09:12.467951 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 12:09:12 crc kubenswrapper[4715]: E1125 12:09:12.468548 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.40:6443: connect: connection refused" node="crc" Nov 25 12:09:12 crc kubenswrapper[4715]: I1125 12:09:12.582708 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 25 12:09:12 crc kubenswrapper[4715]: E1125 12:09:12.583894 4715 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:12 crc kubenswrapper[4715]: I1125 12:09:12.627926 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:12 crc kubenswrapper[4715]: I1125 12:09:12.631963 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 13:30:53.089401171 +0000 UTC Nov 25 12:09:12 crc kubenswrapper[4715]: I1125 12:09:12.632028 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1129h21m40.457377116s for next certificate rotation Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.628761 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:13 crc kubenswrapper[4715]: E1125 12:09:13.646560 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="3.2s" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.709729 4715 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="e18d1184371d56342b390112ef59d78776329571eabf044ab604806c04acf7b0" exitCode=0 Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.709857 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"e18d1184371d56342b390112ef59d78776329571eabf044ab604806c04acf7b0"} Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.709886 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.711650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.711685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.711696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.712014 4715 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783" exitCode=0 Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.712089 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783"} Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.712236 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.713237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.713265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.713276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:13 crc kubenswrapper[4715]: W1125 12:09:13.713875 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:13 crc kubenswrapper[4715]: E1125 12:09:13.713942 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.714478 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c" exitCode=0 Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.714538 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c"} Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.714576 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.715992 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.716045 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.716059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.717671 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f"} Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.717716 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25"} Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.717734 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637"} Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.719240 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="39c7a80a7f8bf43652f7ed2c89123140f7382ec852049e9425458acec423a0b4" exitCode=0 Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.719300 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"39c7a80a7f8bf43652f7ed2c89123140f7382ec852049e9425458acec423a0b4"} Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.719389 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.720415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.720440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.720451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.720606 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.721459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.721503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:13 crc kubenswrapper[4715]: I1125 12:09:13.721521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:13 crc kubenswrapper[4715]: E1125 12:09:13.778089 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.40:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b3ea36a0948d9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 12:09:10.625265881 +0000 UTC m=+1.132768902,LastTimestamp:2025-11-25 12:09:10.625265881 +0000 UTC m=+1.132768902,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 12:09:14 crc kubenswrapper[4715]: W1125 12:09:14.025575 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:14 crc kubenswrapper[4715]: E1125 12:09:14.025658 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.069330 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.071096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.071147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.071159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.071207 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 12:09:14 crc kubenswrapper[4715]: E1125 12:09:14.071908 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.40:6443: connect: connection refused" node="crc" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.628054 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:14 crc kubenswrapper[4715]: W1125 12:09:14.724441 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:14 crc kubenswrapper[4715]: E1125 12:09:14.724537 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.726179 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040"} Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.726446 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.727512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.727555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.727569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.729467 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="39f417bb62facafa423eac71cfdc98e439ca08948cae893d49dda30dddbe57e2" exitCode=0 Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.729523 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"39f417bb62facafa423eac71cfdc98e439ca08948cae893d49dda30dddbe57e2"} Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.729638 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.730518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.730542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.730552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.733428 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ae9d84c6787b11b098790ccb9a5ab681b38105500e922ed877f81d7d782adfb9"} Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.733501 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.734789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.734824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.734843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.736417 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451"} Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.736448 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04"} Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.736461 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1"} Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.736570 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.738622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.738673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.738686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.743362 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd"} Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.743471 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae"} Nov 25 12:09:14 crc kubenswrapper[4715]: I1125 12:09:14.743487 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1"} Nov 25 12:09:15 crc kubenswrapper[4715]: W1125 12:09:15.007139 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:15 crc kubenswrapper[4715]: E1125 12:09:15.007235 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.40:6443: connect: connection refused" logger="UnhandledError" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.532865 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.628631 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.40:6443: connect: connection refused Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.750572 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5e68976a8f880cce69dbbc579d5c9066202b43193ec631fb6817f5a2285ba2e4"} Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.750645 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728"} Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.750762 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.752018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.752052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.752064 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.753398 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="258869dd8353aa6544618e6c080fba6d21023f16576ea92dd55d74faf180fa7a" exitCode=0 Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.753520 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.753541 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.753553 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.753521 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"258869dd8353aa6544618e6c080fba6d21023f16576ea92dd55d74faf180fa7a"} Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.754800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.755279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.755303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:15 crc kubenswrapper[4715]: I1125 12:09:15.755313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.597826 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.695639 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.757958 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.760430 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5e68976a8f880cce69dbbc579d5c9066202b43193ec631fb6817f5a2285ba2e4" exitCode=255 Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.760529 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5e68976a8f880cce69dbbc579d5c9066202b43193ec631fb6817f5a2285ba2e4"} Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.760702 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.762128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.762166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.762177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.762857 4715 scope.go:117] "RemoveContainer" containerID="5e68976a8f880cce69dbbc579d5c9066202b43193ec631fb6817f5a2285ba2e4" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.771427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fde5268d02a6206f73d1a64b77d7145092e59758bc01c80e096b9e12d3f78d3c"} Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.771496 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"df6d9273ca8652c52189a435aa4b06d724fa18dbca2b78141e63e695b7f6f3d9"} Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.771523 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5ade9a469a8f08e584b0d95b466bd612154a082b1487eebb198739c1b198c595"} Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.771502 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.771627 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.773645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.773753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.773768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.773821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.773898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.773916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:16 crc kubenswrapper[4715]: I1125 12:09:16.873329 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.272920 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.274501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.274541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.274556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.274582 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.739399 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.775643 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.777527 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84"} Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.777672 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.778150 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.778598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.778626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.778637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.784960 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b4fe863afe95d6e4d7bfbd829de3bd73e2fc041f2cc75aae83922cd063a102fd"} Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.785014 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b42e0d2a865c7b7b7ae6efedbcbe6f9ca83e8f71c2334caf3c24a587acd1a574"} Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.785055 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.785966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.785999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:17 crc kubenswrapper[4715]: I1125 12:09:17.786012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:18 crc kubenswrapper[4715]: I1125 12:09:18.788476 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:18 crc kubenswrapper[4715]: I1125 12:09:18.788582 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:18 crc kubenswrapper[4715]: I1125 12:09:18.789560 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:18 crc kubenswrapper[4715]: I1125 12:09:18.790420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:18 crc kubenswrapper[4715]: I1125 12:09:18.790476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:18 crc kubenswrapper[4715]: I1125 12:09:18.790491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:18 crc kubenswrapper[4715]: I1125 12:09:18.790622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:18 crc kubenswrapper[4715]: I1125 12:09:18.790659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:18 crc kubenswrapper[4715]: I1125 12:09:18.790672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:19 crc kubenswrapper[4715]: I1125 12:09:19.791958 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:19 crc kubenswrapper[4715]: I1125 12:09:19.793149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:19 crc kubenswrapper[4715]: I1125 12:09:19.793245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:19 crc kubenswrapper[4715]: I1125 12:09:19.793266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:19 crc kubenswrapper[4715]: I1125 12:09:19.888397 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 12:09:19 crc kubenswrapper[4715]: I1125 12:09:19.888673 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:19 crc kubenswrapper[4715]: I1125 12:09:19.890106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:19 crc kubenswrapper[4715]: I1125 12:09:19.890161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:19 crc kubenswrapper[4715]: I1125 12:09:19.890176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:20 crc kubenswrapper[4715]: E1125 12:09:20.964942 4715 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 12:09:21 crc kubenswrapper[4715]: I1125 12:09:21.689700 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:21 crc kubenswrapper[4715]: I1125 12:09:21.689897 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:21 crc kubenswrapper[4715]: I1125 12:09:21.691273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:21 crc kubenswrapper[4715]: I1125 12:09:21.691317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:21 crc kubenswrapper[4715]: I1125 12:09:21.691331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.404098 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.404375 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.405834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.405935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.405948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.408440 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.799277 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.799469 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.800833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.800889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.800905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:22 crc kubenswrapper[4715]: I1125 12:09:22.803776 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:23 crc kubenswrapper[4715]: I1125 12:09:23.802514 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:23 crc kubenswrapper[4715]: I1125 12:09:23.804099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:23 crc kubenswrapper[4715]: I1125 12:09:23.804131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:23 crc kubenswrapper[4715]: I1125 12:09:23.804141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:24 crc kubenswrapper[4715]: I1125 12:09:24.690382 4715 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 12:09:24 crc kubenswrapper[4715]: I1125 12:09:24.690475 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 12:09:24 crc kubenswrapper[4715]: I1125 12:09:24.804412 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:24 crc kubenswrapper[4715]: I1125 12:09:24.805761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:24 crc kubenswrapper[4715]: I1125 12:09:24.805811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:24 crc kubenswrapper[4715]: I1125 12:09:24.805835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.494989 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.495257 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.496426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.496463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.496474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.542100 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.598988 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded" start-of-body= Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.599093 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.628583 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.809450 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.810868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.811002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.811094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:26 crc kubenswrapper[4715]: I1125 12:09:26.822581 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 12:09:26 crc kubenswrapper[4715]: E1125 12:09:26.848450 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 12:09:26 crc kubenswrapper[4715]: E1125 12:09:26.875803 4715 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 12:09:27 crc kubenswrapper[4715]: I1125 12:09:27.080153 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 12:09:27 crc kubenswrapper[4715]: I1125 12:09:27.080621 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 12:09:27 crc kubenswrapper[4715]: I1125 12:09:27.833923 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:27 crc kubenswrapper[4715]: I1125 12:09:27.835401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:27 crc kubenswrapper[4715]: I1125 12:09:27.835436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:27 crc kubenswrapper[4715]: I1125 12:09:27.835445 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:30 crc kubenswrapper[4715]: E1125 12:09:30.965223 4715 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.604282 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.604503 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.604937 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.605027 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.606388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.606435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.606449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.608783 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.842927 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.843511 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.843589 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.844007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.844078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:31 crc kubenswrapper[4715]: I1125 12:09:31.844094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.079696 4715 trace.go:236] Trace[1510789313]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 12:09:19.370) (total time: 12709ms): Nov 25 12:09:32 crc kubenswrapper[4715]: Trace[1510789313]: ---"Objects listed" error: 12709ms (12:09:32.079) Nov 25 12:09:32 crc kubenswrapper[4715]: Trace[1510789313]: [12.70935852s] [12.70935852s] END Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.079898 4715 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.080466 4715 trace.go:236] Trace[199269566]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 12:09:19.831) (total time: 12248ms): Nov 25 12:09:32 crc kubenswrapper[4715]: Trace[199269566]: ---"Objects listed" error: 12248ms (12:09:32.080) Nov 25 12:09:32 crc kubenswrapper[4715]: Trace[199269566]: [12.248486628s] [12.248486628s] END Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.080520 4715 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.083898 4715 trace.go:236] Trace[740018253]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 12:09:19.632) (total time: 12451ms): Nov 25 12:09:32 crc kubenswrapper[4715]: Trace[740018253]: ---"Objects listed" error: 12451ms (12:09:32.083) Nov 25 12:09:32 crc kubenswrapper[4715]: Trace[740018253]: [12.45153603s] [12.45153603s] END Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.083966 4715 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.085550 4715 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.087251 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.089970 4715 trace.go:236] Trace[15055481]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 12:09:19.570) (total time: 12519ms): Nov 25 12:09:32 crc kubenswrapper[4715]: Trace[15055481]: ---"Objects listed" error: 12519ms (12:09:32.089) Nov 25 12:09:32 crc kubenswrapper[4715]: Trace[15055481]: [12.51958261s] [12.51958261s] END Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.090021 4715 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.529731 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:32892->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.529844 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:32892->192.168.126.11:17697: read: connection reset by peer" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.618809 4715 apiserver.go:52] "Watching apiserver" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.623337 4715 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.623965 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.624565 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.624723 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.624784 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.624936 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.625316 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.625165 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.625219 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.625135 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.626055 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.626389 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.627217 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.627518 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.628293 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.628690 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.628854 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.628697 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.629667 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.629722 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.630722 4715 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.659070 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.672538 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.684148 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.689507 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.689643 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.689733 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.689880 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690013 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690029 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690047 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690159 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690221 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690259 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690298 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690331 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690367 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690404 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690439 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690479 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690512 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690585 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690620 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690653 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690691 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690728 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690870 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690885 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690945 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.690982 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691016 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691057 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691064 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691089 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691244 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691289 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691312 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691335 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691355 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691383 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691415 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691447 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691476 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691550 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691576 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691603 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691626 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691682 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691704 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691722 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691741 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691762 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691785 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691804 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691824 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691843 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691862 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691879 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691901 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691918 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691936 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691953 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691975 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691992 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692015 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692034 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692050 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692070 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692114 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692132 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692149 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692166 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692201 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692224 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692241 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692262 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692282 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692299 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692316 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692338 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692358 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692381 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692400 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692418 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692436 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692456 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692474 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692492 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692510 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692571 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692597 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692623 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692643 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692661 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692679 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692701 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692721 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692744 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692762 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692782 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692802 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692822 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692843 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692881 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692909 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692927 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692946 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692964 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692982 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692999 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693017 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693034 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693051 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693068 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693086 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693108 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693124 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693143 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693161 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693178 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693212 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691254 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693238 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691568 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691698 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691935 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.691943 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692226 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692341 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692474 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692497 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692611 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.692838 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693037 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693127 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693213 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693371 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693420 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693454 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693705 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693776 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693868 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693964 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694046 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694145 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694247 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694339 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694435 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694514 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694521 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694542 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694845 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.694860 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.695127 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.695141 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.695248 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.695261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696071 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.693230 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696215 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696243 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696273 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696306 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696336 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696362 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696415 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696450 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696483 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696507 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696531 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696566 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696596 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696621 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696647 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696673 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696701 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696736 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696791 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696815 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696841 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696865 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696886 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696905 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696926 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696943 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696961 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696983 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697002 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697021 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697043 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697065 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697088 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697106 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697123 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697141 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697158 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697176 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697222 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697267 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697287 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697305 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697321 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697339 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697357 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697375 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697393 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697410 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697429 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697447 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697464 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697483 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697506 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697526 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697548 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697571 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697593 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697615 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697640 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697665 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697690 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697715 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697745 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697774 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697800 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697825 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697850 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697873 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697896 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697917 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697939 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697962 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697987 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698011 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698037 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698154 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698195 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698220 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698244 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698308 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698341 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698369 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698393 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698425 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698455 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698480 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698506 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698532 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698559 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698612 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698636 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698661 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698754 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698768 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698778 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698790 4715 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698799 4715 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698810 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698859 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698871 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698887 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698901 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698912 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698924 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698938 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698952 4715 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698964 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698978 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698993 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699006 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699019 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699031 4715 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699043 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699053 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699066 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699079 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699092 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699109 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699122 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699135 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699146 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699157 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699166 4715 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699175 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699208 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699223 4715 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699237 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699250 4715 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699261 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699271 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699280 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699291 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699303 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699312 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.700461 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.701663 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696303 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696311 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696311 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696606 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696619 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696654 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696803 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696912 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696960 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696986 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.696994 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697477 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697450 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697728 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697826 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697907 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.697935 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698048 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698146 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698238 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698437 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698466 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698487 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.698600 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699071 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699363 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.699372 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.700335 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.700386 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.700506 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.700738 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.700769 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.700789 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.700956 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.701429 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.703590 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:33.203563213 +0000 UTC m=+23.711066244 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.703632 4715 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.703744 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.703925 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704015 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704123 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704164 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704209 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704414 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704541 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704546 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704706 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704916 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.704930 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.705006 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.705067 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.705289 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.705536 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.705571 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.705955 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.706034 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.706293 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.707163 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.707312 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.707524 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.707584 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.707607 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.707633 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.707893 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.707976 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.708128 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.708395 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.708536 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.708751 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.708874 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.709722 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.709817 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.711918 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.713143 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.713277 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.713375 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.713399 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.714844 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.715049 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.715657 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.716101 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.716100 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.716495 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.716602 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:33.216573088 +0000 UTC m=+23.724076119 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.716985 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:09:33.216968958 +0000 UTC m=+23.724471979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.718635 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.719001 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.719010 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.719129 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.720173 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.720334 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.718924 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.720433 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.702568 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.702442 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.702812 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.702879 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.703234 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.703298 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.720644 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.720743 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.720944 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.721009 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.721131 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.721469 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.721513 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.721553 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.721703 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.721724 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.721989 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.722273 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.722591 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.722775 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.722945 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.722993 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.723386 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.723428 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.723517 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.722703 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.723547 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.723923 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.723960 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.724176 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.724299 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.724315 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.724521 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.724538 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.724810 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.724917 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.724929 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.725240 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.702407 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.725635 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.725668 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.725969 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.725979 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.725430 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.726340 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.726588 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.726473 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.726678 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.726800 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.726792 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.726988 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.727298 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.727541 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.727589 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.728457 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.729317 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.729489 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.730178 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.730419 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.730673 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.733414 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.733444 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.733442 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.733459 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.733599 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.733784 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:33.23375172 +0000 UTC m=+23.741254921 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.733416 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.734056 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.734110 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.734403 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:33.234351145 +0000 UTC m=+23.741854166 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.735509 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.736295 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.736882 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.737417 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.737495 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.738334 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.743036 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.743916 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.744492 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.745624 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.752268 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.754380 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.754399 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.754506 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.755808 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.758630 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.760833 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.761756 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.763761 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.765754 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.765909 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.766750 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.766909 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.767669 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.769345 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.769359 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.769969 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.770451 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.771632 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.772379 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.773537 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.773998 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.774727 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.776587 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.777281 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.778731 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.779287 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.780377 4715 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.780506 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.782466 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.783084 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.784073 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.784556 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.785737 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.786652 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.787572 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.788255 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.789342 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.789856 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.790849 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.791459 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.792548 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.793041 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.793994 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.794518 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.795652 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.796139 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.797047 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.797562 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.798161 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.799221 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.799778 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800564 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800634 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800770 4715 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800792 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800808 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800821 4715 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800834 4715 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800847 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800859 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800871 4715 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800883 4715 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800895 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800907 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800921 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800934 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800968 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800981 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.800993 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801004 4715 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801016 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801028 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801038 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801041 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801069 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801080 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801089 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801097 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801106 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801114 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801122 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801131 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801140 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801154 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801163 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801171 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801180 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801209 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801218 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801227 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801236 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801246 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801269 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801277 4715 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801286 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801294 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801303 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801312 4715 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801320 4715 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801328 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801341 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801350 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801360 4715 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801369 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801392 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801391 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801411 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801421 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801430 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801441 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801450 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801462 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801472 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801481 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801491 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801500 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801508 4715 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801516 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801526 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801535 4715 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801544 4715 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801553 4715 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801564 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801572 4715 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801580 4715 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801589 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801598 4715 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801607 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801615 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801623 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801631 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801640 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801649 4715 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801657 4715 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801665 4715 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801673 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801682 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801690 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801703 4715 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801712 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801722 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801730 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801739 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801748 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801757 4715 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801765 4715 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801773 4715 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801782 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801790 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801798 4715 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801806 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801814 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801822 4715 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801830 4715 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801843 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801854 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801862 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801870 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801878 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801886 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801894 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801902 4715 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801911 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801920 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801928 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801936 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801944 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801952 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801960 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801968 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801977 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801985 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.801994 4715 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802002 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802010 4715 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802017 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802025 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802034 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802042 4715 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802051 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802059 4715 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802067 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802074 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802082 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802091 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802099 4715 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802109 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802117 4715 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802125 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802133 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802142 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802150 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802158 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802167 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802175 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802195 4715 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802203 4715 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802212 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802222 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802231 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802239 4715 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802250 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802258 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802267 4715 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802275 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802283 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.802291 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.847265 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.847782 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.849544 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84" exitCode=255 Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.849586 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84"} Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.849643 4715 scope.go:117] "RemoveContainer" containerID="5e68976a8f880cce69dbbc579d5c9066202b43193ec631fb6817f5a2285ba2e4" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.861817 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.862376 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.862735 4715 scope.go:117] "RemoveContainer" containerID="3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84" Nov 25 12:09:32 crc kubenswrapper[4715]: E1125 12:09:32.863159 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.870377 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.877927 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.885475 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.898528 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.909313 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.909468 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.913567 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.921682 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e68976a8f880cce69dbbc579d5c9066202b43193ec631fb6817f5a2285ba2e4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:15Z\\\",\\\"message\\\":\\\"W1125 12:09:15.180505 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1125 12:09:15.181455 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764072555 cert, and key in /tmp/serving-cert-2854598065/serving-signer.crt, /tmp/serving-cert-2854598065/serving-signer.key\\\\nI1125 12:09:15.658893 1 observer_polling.go:159] Starting file observer\\\\nW1125 12:09:15.661693 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 12:09:15.661862 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:15.664931 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2854598065/tls.crt::/tmp/serving-cert-2854598065/tls.key\\\\\\\"\\\\nF1125 12:09:15.935035 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.931614 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.938747 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.940126 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.945805 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.950632 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.953701 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 12:09:32 crc kubenswrapper[4715]: W1125 12:09:32.952829 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-8aa2877057698d02a0c7d7bae97c5d5069f0e10f10460af2d46426d75660e4e2 WatchSource:0}: Error finding container 8aa2877057698d02a0c7d7bae97c5d5069f0e10f10460af2d46426d75660e4e2: Status 404 returned error can't find the container with id 8aa2877057698d02a0c7d7bae97c5d5069f0e10f10460af2d46426d75660e4e2 Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.961430 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.962857 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 12:09:32 crc kubenswrapper[4715]: W1125 12:09:32.968352 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-5d5d1d12b38d9f9560be8a96304e2d6959c7083d80673ba2b4ec9fb54fdfafaa WatchSource:0}: Error finding container 5d5d1d12b38d9f9560be8a96304e2d6959c7083d80673ba2b4ec9fb54fdfafaa: Status 404 returned error can't find the container with id 5d5d1d12b38d9f9560be8a96304e2d6959c7083d80673ba2b4ec9fb54fdfafaa Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.971620 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.984623 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:32 crc kubenswrapper[4715]: I1125 12:09:32.993661 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.004983 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.015975 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e68976a8f880cce69dbbc579d5c9066202b43193ec631fb6817f5a2285ba2e4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:15Z\\\",\\\"message\\\":\\\"W1125 12:09:15.180505 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1125 12:09:15.181455 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764072555 cert, and key in /tmp/serving-cert-2854598065/serving-signer.crt, /tmp/serving-cert-2854598065/serving-signer.key\\\\nI1125 12:09:15.658893 1 observer_polling.go:159] Starting file observer\\\\nW1125 12:09:15.661693 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 12:09:15.661862 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:15.664931 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2854598065/tls.crt::/tmp/serving-cert-2854598065/tls.key\\\\\\\"\\\\nF1125 12:09:15.935035 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.027991 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.038098 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.048339 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.059351 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.069091 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.207500 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.207682 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.207779 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:34.207750526 +0000 UTC m=+24.715253547 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.308175 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308302 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:09:34.308274762 +0000 UTC m=+24.815777783 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.308341 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.308380 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.308409 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308508 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308524 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308541 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308555 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308568 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:34.308557989 +0000 UTC m=+24.816061010 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308591 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:34.30858138 +0000 UTC m=+24.816084401 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308648 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308694 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308713 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.308808 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:34.308776975 +0000 UTC m=+24.816279996 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.703318 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.854325 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5d5d1d12b38d9f9560be8a96304e2d6959c7083d80673ba2b4ec9fb54fdfafaa"} Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.856975 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92"} Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.857005 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1"} Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.857017 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"627c1862cb7ba07b1d6cc714cbf69c9f77efe7745ad44dadeb17f03da3e367ec"} Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.858933 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43"} Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.858959 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8aa2877057698d02a0c7d7bae97c5d5069f0e10f10460af2d46426d75660e4e2"} Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.861063 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.863823 4715 scope.go:117] "RemoveContainer" containerID="3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84" Nov 25 12:09:33 crc kubenswrapper[4715]: E1125 12:09:33.863948 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.873083 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.884099 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.898866 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e68976a8f880cce69dbbc579d5c9066202b43193ec631fb6817f5a2285ba2e4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:15Z\\\",\\\"message\\\":\\\"W1125 12:09:15.180505 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1125 12:09:15.181455 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764072555 cert, and key in /tmp/serving-cert-2854598065/serving-signer.crt, /tmp/serving-cert-2854598065/serving-signer.key\\\\nI1125 12:09:15.658893 1 observer_polling.go:159] Starting file observer\\\\nW1125 12:09:15.661693 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1125 12:09:15.661862 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:15.664931 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2854598065/tls.crt::/tmp/serving-cert-2854598065/tls.key\\\\\\\"\\\\nF1125 12:09:15.935035 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.912393 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.925937 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.937779 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.948241 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.958919 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.970014 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.980291 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:33 crc kubenswrapper[4715]: I1125 12:09:33.990865 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.001324 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.011341 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.022927 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.032799 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.046395 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.218542 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.218664 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.218738 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:36.218717188 +0000 UTC m=+26.726220209 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.320034 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320148 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:09:36.320126507 +0000 UTC m=+26.827629518 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.320333 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.320402 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320492 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320565 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320585 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320597 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320599 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:36.320582969 +0000 UTC m=+26.828086000 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320642 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:36.3206308 +0000 UTC m=+26.828133831 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320846 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320905 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.320920 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.321003 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:36.32097143 +0000 UTC m=+26.828474451 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.320433 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.693390 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.693472 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.693615 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.693654 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.693806 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.694008 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.698049 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.698915 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.699851 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.700699 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.701439 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.702271 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 12:09:34 crc kubenswrapper[4715]: I1125 12:09:34.868601 4715 scope.go:117] "RemoveContainer" containerID="3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84" Nov 25 12:09:34 crc kubenswrapper[4715]: E1125 12:09:34.868844 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 12:09:35 crc kubenswrapper[4715]: I1125 12:09:35.577277 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 25 12:09:35 crc kubenswrapper[4715]: I1125 12:09:35.601322 4715 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 25 12:09:35 crc kubenswrapper[4715]: I1125 12:09:35.873074 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7"} Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.243605 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.243736 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.243814 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:40.243791814 +0000 UTC m=+30.751294835 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.344693 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.344822 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.344886 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:09:40.344852284 +0000 UTC m=+30.852355305 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.344996 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.345084 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:40.34506048 +0000 UTC m=+30.852563701 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.345116 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.345135 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.345149 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.345213 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:40.345205123 +0000 UTC m=+30.852708144 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.344855 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.345662 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.345767 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.345888 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.345903 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.345941 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:40.345932712 +0000 UTC m=+30.853435723 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.478528 4715 csr.go:261] certificate signing request csr-tnd5k is approved, waiting to be issued Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.497469 4715 csr.go:257] certificate signing request csr-tnd5k is issued Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.542848 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-qw7ss"] Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.543260 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qw7ss" Nov 25 12:09:36 crc kubenswrapper[4715]: W1125 12:09:36.545363 4715 reflector.go:561] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": failed to list *v1.Secret: secrets "node-resolver-dockercfg-kz9s7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.545405 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"node-resolver-dockercfg-kz9s7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"node-resolver-dockercfg-kz9s7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 12:09:36 crc kubenswrapper[4715]: W1125 12:09:36.545846 4715 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.545871 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 12:09:36 crc kubenswrapper[4715]: W1125 12:09:36.545934 4715 reflector.go:561] object-"openshift-dns"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.545948 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.599793 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.622310 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.651565 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxcrw\" (UniqueName: \"kubernetes.io/projected/e6ba7789-d2b7-46eb-97fd-0f61f98e8e15-kube-api-access-kxcrw\") pod \"node-resolver-qw7ss\" (UID: \"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\") " pod="openshift-dns/node-resolver-qw7ss" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.651637 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e6ba7789-d2b7-46eb-97fd-0f61f98e8e15-hosts-file\") pod \"node-resolver-qw7ss\" (UID: \"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\") " pod="openshift-dns/node-resolver-qw7ss" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.664080 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.693343 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.693372 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.693397 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.693496 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.693585 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:36 crc kubenswrapper[4715]: E1125 12:09:36.693735 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.697250 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.716402 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.736429 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.752475 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxcrw\" (UniqueName: \"kubernetes.io/projected/e6ba7789-d2b7-46eb-97fd-0f61f98e8e15-kube-api-access-kxcrw\") pod \"node-resolver-qw7ss\" (UID: \"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\") " pod="openshift-dns/node-resolver-qw7ss" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.752540 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e6ba7789-d2b7-46eb-97fd-0f61f98e8e15-hosts-file\") pod \"node-resolver-qw7ss\" (UID: \"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\") " pod="openshift-dns/node-resolver-qw7ss" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.752627 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e6ba7789-d2b7-46eb-97fd-0f61f98e8e15-hosts-file\") pod \"node-resolver-qw7ss\" (UID: \"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\") " pod="openshift-dns/node-resolver-qw7ss" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.755983 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.770512 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.780823 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.891871 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.908850 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.927457 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.939380 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.955470 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.970740 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:36 crc kubenswrapper[4715]: I1125 12:09:36.985117 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.001405 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:36Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.016614 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.043520 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-5mzr5"] Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.043996 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.047118 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.047202 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.049918 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.051286 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.052678 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.069265 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.104387 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.132131 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.155876 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-os-release\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.155941 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-multus-socket-dir-parent\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.155970 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-var-lib-cni-multus\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.155997 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-var-lib-kubelet\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156126 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6371ac18-2361-43bb-b474-32bf53713cf5-multus-daemon-config\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156226 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-etc-kubernetes\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156246 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw5xd\" (UniqueName: \"kubernetes.io/projected/6371ac18-2361-43bb-b474-32bf53713cf5-kube-api-access-rw5xd\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156286 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-multus-cni-dir\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156312 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-var-lib-cni-bin\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156489 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-system-cni-dir\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156584 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-run-k8s-cni-cncf-io\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156613 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-run-multus-certs\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156720 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-cnibin\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156787 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-run-netns\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156842 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-hostroot\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156876 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-multus-conf-dir\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.156934 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6371ac18-2361-43bb-b474-32bf53713cf5-cni-binary-copy\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.161346 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.176248 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.188388 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.203968 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.215403 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.234597 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.251077 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257395 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6371ac18-2361-43bb-b474-32bf53713cf5-multus-daemon-config\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257452 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-etc-kubernetes\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257479 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw5xd\" (UniqueName: \"kubernetes.io/projected/6371ac18-2361-43bb-b474-32bf53713cf5-kube-api-access-rw5xd\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257516 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-multus-cni-dir\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257545 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-var-lib-cni-bin\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-etc-kubernetes\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257609 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-system-cni-dir\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257674 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-var-lib-cni-bin\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257709 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-run-k8s-cni-cncf-io\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257734 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-run-multus-certs\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257757 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-multus-cni-dir\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257784 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-cnibin\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257796 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-run-k8s-cni-cncf-io\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257819 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-run-multus-certs\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257832 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6371ac18-2361-43bb-b474-32bf53713cf5-cni-binary-copy\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257761 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-system-cni-dir\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257859 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-run-netns\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-cnibin\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257882 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-hostroot\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257910 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-hostroot\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257953 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-multus-conf-dir\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257910 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-run-netns\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.257989 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-os-release\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258011 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-multus-socket-dir-parent\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258031 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-var-lib-cni-multus\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258043 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-multus-conf-dir\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258083 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-var-lib-kubelet\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258085 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6371ac18-2361-43bb-b474-32bf53713cf5-multus-daemon-config\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258125 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-var-lib-kubelet\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258129 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-host-var-lib-cni-multus\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-multus-socket-dir-parent\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258276 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6371ac18-2361-43bb-b474-32bf53713cf5-os-release\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.258361 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6371ac18-2361-43bb-b474-32bf53713cf5-cni-binary-copy\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.276747 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw5xd\" (UniqueName: \"kubernetes.io/projected/6371ac18-2361-43bb-b474-32bf53713cf5-kube-api-access-rw5xd\") pod \"multus-5mzr5\" (UID: \"6371ac18-2361-43bb-b474-32bf53713cf5\") " pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.358806 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5mzr5" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.460722 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-dk9f9"] Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.461125 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.463864 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.463926 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.464430 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.464830 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.465220 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.466394 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tzl95"] Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.467264 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.468761 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-5fwzm"] Nov 25 12:09:37 crc kubenswrapper[4715]: W1125 12:09:37.469658 4715 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-pwtwl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Nov 25 12:09:37 crc kubenswrapper[4715]: E1125 12:09:37.469705 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-pwtwl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-node-dockercfg-pwtwl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 12:09:37 crc kubenswrapper[4715]: W1125 12:09:37.469753 4715 reflector.go:561] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Nov 25 12:09:37 crc kubenswrapper[4715]: E1125 12:09:37.469764 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 12:09:37 crc kubenswrapper[4715]: W1125 12:09:37.475025 4715 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": failed to list *v1.ConfigMap: configmaps "ovnkube-script-lib" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Nov 25 12:09:37 crc kubenswrapper[4715]: E1125 12:09:37.475134 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnkube-script-lib\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.475210 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.475745 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.476070 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.476565 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.479067 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.481176 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.481533 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.503406 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.513264 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-25 12:04:36 +0000 UTC, rotation deadline is 2026-08-14 22:35:18.780733136 +0000 UTC Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.513391 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6298h25m41.267346935s for next certificate rotation Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.523370 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.543819 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.556216 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561416 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-openvswitch\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561446 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-netd\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561462 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-kubelet\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561481 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpqs9\" (UniqueName: \"kubernetes.io/projected/02385935-e23a-4708-b57f-fd393d05651b-kube-api-access-kpqs9\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561508 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/51676837-adc4-4424-b527-920a6528b6a2-proxy-tls\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561525 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-ovn-kubernetes\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561540 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561564 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qn9f\" (UniqueName: \"kubernetes.io/projected/51676837-adc4-4424-b527-920a6528b6a2-kube-api-access-7qn9f\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561579 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-node-log\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561665 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-ovn\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561742 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-script-lib\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/51676837-adc4-4424-b527-920a6528b6a2-rootfs\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561933 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-systemd-units\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.561968 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-etc-openvswitch\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562017 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-env-overrides\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562049 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/51676837-adc4-4424-b527-920a6528b6a2-mcd-auth-proxy-config\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562098 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/02385935-e23a-4708-b57f-fd393d05651b-ovn-node-metrics-cert\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562321 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-slash\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562539 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-netns\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562565 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-bin\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562752 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-log-socket\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562779 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-config\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562800 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-systemd\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.562821 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-var-lib-openvswitch\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.572423 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.588326 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.604089 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.631731 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.648555 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.660212 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663695 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-kubelet\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663744 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpqs9\" (UniqueName: \"kubernetes.io/projected/02385935-e23a-4708-b57f-fd393d05651b-kube-api-access-kpqs9\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663769 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/51676837-adc4-4424-b527-920a6528b6a2-proxy-tls\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663794 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-ovn-kubernetes\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663813 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663838 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-os-release\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663855 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663880 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qn9f\" (UniqueName: \"kubernetes.io/projected/51676837-adc4-4424-b527-920a6528b6a2-kube-api-access-7qn9f\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663880 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-kubelet\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663936 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-node-log\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663959 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-ovn-kubernetes\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663967 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.663896 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-node-log\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664063 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-ovn\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664088 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-script-lib\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664114 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-etc-openvswitch\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664109 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-ovn\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664147 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-system-cni-dir\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664165 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-etc-openvswitch\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664176 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/51676837-adc4-4424-b527-920a6528b6a2-rootfs\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664229 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/51676837-adc4-4424-b527-920a6528b6a2-rootfs\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664227 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-systemd-units\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664250 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-systemd-units\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664270 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-env-overrides\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664300 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/02385935-e23a-4708-b57f-fd393d05651b-ovn-node-metrics-cert\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664386 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-cnibin\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664420 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/51676837-adc4-4424-b527-920a6528b6a2-mcd-auth-proxy-config\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664459 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9922\" (UniqueName: \"kubernetes.io/projected/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-kube-api-access-h9922\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664487 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-bin\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664514 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-slash\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664537 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-netns\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664566 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-log-socket\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664588 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-config\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664613 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-systemd\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664649 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-var-lib-openvswitch\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664672 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664697 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664725 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-openvswitch\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664744 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-netd\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664804 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-netd\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664852 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-bin\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664880 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-slash\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664907 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-netns\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.664933 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-log-socket\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.665035 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-systemd\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.665053 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-env-overrides\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.665067 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/51676837-adc4-4424-b527-920a6528b6a2-mcd-auth-proxy-config\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.665079 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-openvswitch\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.665096 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-var-lib-openvswitch\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.665533 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-config\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.668744 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/02385935-e23a-4708-b57f-fd393d05651b-ovn-node-metrics-cert\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.668922 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/51676837-adc4-4424-b527-920a6528b6a2-proxy-tls\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.671855 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.681387 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qn9f\" (UniqueName: \"kubernetes.io/projected/51676837-adc4-4424-b527-920a6528b6a2-kube-api-access-7qn9f\") pod \"machine-config-daemon-dk9f9\" (UID: \"51676837-adc4-4424-b527-920a6528b6a2\") " pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.686605 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.701425 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.718240 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.736893 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.759077 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.765145 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-cnibin\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.765226 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9922\" (UniqueName: \"kubernetes.io/projected/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-kube-api-access-h9922\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.765269 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.765295 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.765360 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-os-release\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.765382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.765423 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-system-cni-dir\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.765496 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-system-cni-dir\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.765545 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-cnibin\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.766112 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-os-release\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.766244 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.767066 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.767222 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: E1125 12:09:37.768318 4715 projected.go:288] Couldn't get configMap openshift-dns/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.780355 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.782981 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.786426 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9922\" (UniqueName: \"kubernetes.io/projected/f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c-kube-api-access-h9922\") pod \"multus-additional-cni-plugins-5fwzm\" (UID: \"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\") " pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.790974 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 12:09:37 crc kubenswrapper[4715]: W1125 12:09:37.794505 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51676837_adc4_4424_b527_920a6528b6a2.slice/crio-e0e144442ba24a2cfe586a56fbcb9caa322defdf52152a9560777526600856ef WatchSource:0}: Error finding container e0e144442ba24a2cfe586a56fbcb9caa322defdf52152a9560777526600856ef: Status 404 returned error can't find the container with id e0e144442ba24a2cfe586a56fbcb9caa322defdf52152a9560777526600856ef Nov 25 12:09:37 crc kubenswrapper[4715]: E1125 12:09:37.798582 4715 projected.go:194] Error preparing data for projected volume kube-api-access-kxcrw for pod openshift-dns/node-resolver-qw7ss: failed to sync configmap cache: timed out waiting for the condition Nov 25 12:09:37 crc kubenswrapper[4715]: E1125 12:09:37.798793 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e6ba7789-d2b7-46eb-97fd-0f61f98e8e15-kube-api-access-kxcrw podName:e6ba7789-d2b7-46eb-97fd-0f61f98e8e15 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:38.298762974 +0000 UTC m=+28.806266005 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-kxcrw" (UniqueName: "kubernetes.io/projected/e6ba7789-d2b7-46eb-97fd-0f61f98e8e15-kube-api-access-kxcrw") pod "node-resolver-qw7ss" (UID: "e6ba7789-d2b7-46eb-97fd-0f61f98e8e15") : failed to sync configmap cache: timed out waiting for the condition Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.801705 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.804655 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.817331 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.821561 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 12:09:37 crc kubenswrapper[4715]: W1125 12:09:37.823256 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8ba0cf3_8a5a_4282_a6f0_c5119a7cff2c.slice/crio-f1f1b8e235814f755cf180343b96bdf734115ad6514eda225c32ee944be9b5b3 WatchSource:0}: Error finding container f1f1b8e235814f755cf180343b96bdf734115ad6514eda225c32ee944be9b5b3: Status 404 returned error can't find the container with id f1f1b8e235814f755cf180343b96bdf734115ad6514eda225c32ee944be9b5b3 Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.832229 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.846140 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.856571 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.880205 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.883609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"e0e144442ba24a2cfe586a56fbcb9caa322defdf52152a9560777526600856ef"} Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.885621 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerStarted","Data":"f1f1b8e235814f755cf180343b96bdf734115ad6514eda225c32ee944be9b5b3"} Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.892705 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5mzr5" event={"ID":"6371ac18-2361-43bb-b474-32bf53713cf5","Type":"ContainerStarted","Data":"eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798"} Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.892756 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5mzr5" event={"ID":"6371ac18-2361-43bb-b474-32bf53713cf5","Type":"ContainerStarted","Data":"11d37a61f200516433e951a66d8c2592b0051e7ccd0626d58d53297a39e8f908"} Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.899976 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.912661 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.925864 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.939011 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.949387 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.956902 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.961665 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.979426 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:37 crc kubenswrapper[4715]: I1125 12:09:37.993350 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:37Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.008321 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.021405 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.033295 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.044262 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.057213 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.074563 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.375737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxcrw\" (UniqueName: \"kubernetes.io/projected/e6ba7789-d2b7-46eb-97fd-0f61f98e8e15-kube-api-access-kxcrw\") pod \"node-resolver-qw7ss\" (UID: \"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\") " pod="openshift-dns/node-resolver-qw7ss" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.380013 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxcrw\" (UniqueName: \"kubernetes.io/projected/e6ba7789-d2b7-46eb-97fd-0f61f98e8e15-kube-api-access-kxcrw\") pod \"node-resolver-qw7ss\" (UID: \"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\") " pod="openshift-dns/node-resolver-qw7ss" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.487698 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.489693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.489750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.489763 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.489861 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.498509 4715 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.500224 4715 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.501453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.501496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.501511 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.501530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.501545 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:38Z","lastTransitionTime":"2025-11-25T12:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.522119 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.527842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.527916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.527935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.527964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.527982 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:38Z","lastTransitionTime":"2025-11-25T12:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.543171 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.546982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.547067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.547076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.547092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.547105 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:38Z","lastTransitionTime":"2025-11-25T12:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.559470 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.563373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.563431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.563447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.563469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.563481 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:38Z","lastTransitionTime":"2025-11-25T12:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.575407 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.578875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.578923 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.578936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.578953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.578965 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:38Z","lastTransitionTime":"2025-11-25T12:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.590721 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.590895 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.592580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.592616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.592625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.592642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.592652 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:38Z","lastTransitionTime":"2025-11-25T12:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.655780 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qw7ss" Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.664338 4715 configmap.go:193] Couldn't get configMap openshift-ovn-kubernetes/ovnkube-script-lib: failed to sync configmap cache: timed out waiting for the condition Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.664464 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-script-lib podName:02385935-e23a-4708-b57f-fd393d05651b nodeName:}" failed. No retries permitted until 2025-11-25 12:09:39.164440008 +0000 UTC m=+29.671943029 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovnkube-script-lib" (UniqueName: "kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-script-lib") pod "ovnkube-node-tzl95" (UID: "02385935-e23a-4708-b57f-fd393d05651b") : failed to sync configmap cache: timed out waiting for the condition Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.678891 4715 projected.go:288] Couldn't get configMap openshift-ovn-kubernetes/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.679333 4715 projected.go:194] Error preparing data for projected volume kube-api-access-kpqs9 for pod openshift-ovn-kubernetes/ovnkube-node-tzl95: failed to sync configmap cache: timed out waiting for the condition Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.679423 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/02385935-e23a-4708-b57f-fd393d05651b-kube-api-access-kpqs9 podName:02385935-e23a-4708-b57f-fd393d05651b nodeName:}" failed. No retries permitted until 2025-11-25 12:09:39.179391292 +0000 UTC m=+29.686894313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-kpqs9" (UniqueName: "kubernetes.io/projected/02385935-e23a-4708-b57f-fd393d05651b-kube-api-access-kpqs9") pod "ovnkube-node-tzl95" (UID: "02385935-e23a-4708-b57f-fd393d05651b") : failed to sync configmap cache: timed out waiting for the condition Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.688110 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.692934 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.693028 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.693077 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.693211 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.693383 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:38 crc kubenswrapper[4715]: E1125 12:09:38.693460 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.694963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.694997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.695007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.695026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.695040 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:38Z","lastTransitionTime":"2025-11-25T12:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.748963 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.776404 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.799095 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.799133 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.799143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.799161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.799173 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:38Z","lastTransitionTime":"2025-11-25T12:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.897977 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d"} Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.898033 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71"} Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.899450 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qw7ss" event={"ID":"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15","Type":"ContainerStarted","Data":"60300904a553980cec1ee3f83e47ba2c0ef78b5bf2936406074e1dd18c0d4fd3"} Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.900726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.900765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.900779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.900810 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.900823 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:38Z","lastTransitionTime":"2025-11-25T12:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.903241 4715 generic.go:334] "Generic (PLEG): container finished" podID="f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c" containerID="7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af" exitCode=0 Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.903300 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerDied","Data":"7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af"} Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.913645 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.926581 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.940995 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.955562 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-rncj5"] Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.956061 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.956015 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.958761 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.958804 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.961561 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.961664 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.972592 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.982333 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/92f2a401-5b59-48a9-8e15-35d5f62afd66-serviceca\") pod \"node-ca-rncj5\" (UID: \"92f2a401-5b59-48a9-8e15-35d5f62afd66\") " pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.982449 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8ndn\" (UniqueName: \"kubernetes.io/projected/92f2a401-5b59-48a9-8e15-35d5f62afd66-kube-api-access-x8ndn\") pod \"node-ca-rncj5\" (UID: \"92f2a401-5b59-48a9-8e15-35d5f62afd66\") " pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.982520 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92f2a401-5b59-48a9-8e15-35d5f62afd66-host\") pod \"node-ca-rncj5\" (UID: \"92f2a401-5b59-48a9-8e15-35d5f62afd66\") " pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:38 crc kubenswrapper[4715]: I1125 12:09:38.991094 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:38Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.004482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.004518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.004529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.004544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.004555 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.007233 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.021200 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.039080 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.054790 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.070847 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.083643 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92f2a401-5b59-48a9-8e15-35d5f62afd66-host\") pod \"node-ca-rncj5\" (UID: \"92f2a401-5b59-48a9-8e15-35d5f62afd66\") " pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.083713 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/92f2a401-5b59-48a9-8e15-35d5f62afd66-serviceca\") pod \"node-ca-rncj5\" (UID: \"92f2a401-5b59-48a9-8e15-35d5f62afd66\") " pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.083764 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8ndn\" (UniqueName: \"kubernetes.io/projected/92f2a401-5b59-48a9-8e15-35d5f62afd66-kube-api-access-x8ndn\") pod \"node-ca-rncj5\" (UID: \"92f2a401-5b59-48a9-8e15-35d5f62afd66\") " pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.083865 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92f2a401-5b59-48a9-8e15-35d5f62afd66-host\") pod \"node-ca-rncj5\" (UID: \"92f2a401-5b59-48a9-8e15-35d5f62afd66\") " pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.084993 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/92f2a401-5b59-48a9-8e15-35d5f62afd66-serviceca\") pod \"node-ca-rncj5\" (UID: \"92f2a401-5b59-48a9-8e15-35d5f62afd66\") " pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.085005 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.106058 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8ndn\" (UniqueName: \"kubernetes.io/projected/92f2a401-5b59-48a9-8e15-35d5f62afd66-kube-api-access-x8ndn\") pod \"node-ca-rncj5\" (UID: \"92f2a401-5b59-48a9-8e15-35d5f62afd66\") " pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.108706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.108774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.108794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.108822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.108840 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.121475 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.146595 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.184487 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.187095 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpqs9\" (UniqueName: \"kubernetes.io/projected/02385935-e23a-4708-b57f-fd393d05651b-kube-api-access-kpqs9\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.187242 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-script-lib\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.188073 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-script-lib\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.193067 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpqs9\" (UniqueName: \"kubernetes.io/projected/02385935-e23a-4708-b57f-fd393d05651b-kube-api-access-kpqs9\") pod \"ovnkube-node-tzl95\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.207699 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.212424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.212469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.212479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.212495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.212505 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.221571 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.233593 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.245965 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.261497 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.270307 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-rncj5" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.274130 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.292622 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.302062 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.308628 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.317416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.317461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.317471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.317490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.317501 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.319702 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: W1125 12:09:39.325306 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02385935_e23a_4708_b57f_fd393d05651b.slice/crio-4c6fdc92b4855d79d0cf517aa6acd2eec1a1653fa1df8494fa29c67e46a3a965 WatchSource:0}: Error finding container 4c6fdc92b4855d79d0cf517aa6acd2eec1a1653fa1df8494fa29c67e46a3a965: Status 404 returned error can't find the container with id 4c6fdc92b4855d79d0cf517aa6acd2eec1a1653fa1df8494fa29c67e46a3a965 Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.341547 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.363975 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.378621 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.421007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.421065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.421079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.421097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.421109 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.524086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.524138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.524163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.524198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.524215 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.627438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.627473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.627482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.627497 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.627508 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.730165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.730232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.730243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.730262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.730274 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.833101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.833150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.833161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.833205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.833220 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.908914 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-rncj5" event={"ID":"92f2a401-5b59-48a9-8e15-35d5f62afd66","Type":"ContainerStarted","Data":"cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.908986 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-rncj5" event={"ID":"92f2a401-5b59-48a9-8e15-35d5f62afd66","Type":"ContainerStarted","Data":"ecbb4d8084c7692286767accb688fff5276fd2f38cff5595f1568ba944fecdf4"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.911149 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerStarted","Data":"cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.912539 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c" exitCode=0 Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.912602 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.912630 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"4c6fdc92b4855d79d0cf517aa6acd2eec1a1653fa1df8494fa29c67e46a3a965"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.917089 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qw7ss" event={"ID":"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15","Type":"ContainerStarted","Data":"db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.928297 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.936524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.936571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.936592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.936611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.936624 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:39Z","lastTransitionTime":"2025-11-25T12:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.943961 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.955565 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.970009 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.983718 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:39 crc kubenswrapper[4715]: I1125 12:09:39.999131 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.012556 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.023461 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.036459 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.039611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.039644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.039657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.039679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.039693 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.054489 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.083737 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.100281 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.117408 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.132853 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.142149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.142225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.142237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.142258 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.142270 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.147740 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.159043 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.170722 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.180110 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.193012 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.209660 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.221956 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.236022 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.245047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.245086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.245098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.245115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.245128 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.254947 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.268386 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.280787 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.283353 4715 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.300113 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.300242 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.300327 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:48.300299828 +0000 UTC m=+38.807802919 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.307602 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.323979 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.335477 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.347335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.347388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.347428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.347447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.347460 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.401517 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.401726 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.401786 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.401907 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:09:48.401867741 +0000 UTC m=+38.909370762 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.401970 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.402001 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.402015 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.402023 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.402096 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.402116 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.402134 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.402120 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:48.402112758 +0000 UTC m=+38.909615779 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.402241 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.402259 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:48.402232571 +0000 UTC m=+38.909735632 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.402285 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:48.402271632 +0000 UTC m=+38.909774693 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.450996 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.451036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.451046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.451062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.451076 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.553629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.553920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.553929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.553949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.553958 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.657224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.657270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.657282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.657301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.657311 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.693838 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.694003 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.693850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.693850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.694086 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:40 crc kubenswrapper[4715]: E1125 12:09:40.694379 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.707921 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.727653 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.743137 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.755636 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.759536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.759601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.759615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.759636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.759646 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.769535 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.790342 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.807415 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.822699 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.837126 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.857719 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.865776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.865823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.865832 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.865851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.865862 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.881016 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.922465 4715 generic.go:334] "Generic (PLEG): container finished" podID="f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c" containerID="cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2" exitCode=0 Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.922542 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerDied","Data":"cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.925114 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.927715 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.927768 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.946431 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.964397 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.972341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.972381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.972391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.972409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.972420 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:40Z","lastTransitionTime":"2025-11-25T12:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.981521 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:40 crc kubenswrapper[4715]: I1125 12:09:40.996222 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.017070 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.031567 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.042488 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.060674 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.078323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.078381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.078392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.078413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.078424 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:41Z","lastTransitionTime":"2025-11-25T12:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.080479 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.093265 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.106055 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.119621 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.157609 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.181345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.181394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.181405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.181423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.181435 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:41Z","lastTransitionTime":"2025-11-25T12:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.192426 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.239295 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.279520 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.285882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.285924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.285936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.285954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.285964 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:41Z","lastTransitionTime":"2025-11-25T12:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.388770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.388820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.388833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.388861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.388876 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:41Z","lastTransitionTime":"2025-11-25T12:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.491417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.491478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.491498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.491519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.491535 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:41Z","lastTransitionTime":"2025-11-25T12:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.594538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.594596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.594607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.594628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.594641 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:41Z","lastTransitionTime":"2025-11-25T12:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.697083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.697132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.697145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.697165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.697177 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:41Z","lastTransitionTime":"2025-11-25T12:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.801084 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.802258 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.802283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.802299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.802309 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:41Z","lastTransitionTime":"2025-11-25T12:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.905217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.905613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.905719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.905814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.905903 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:41Z","lastTransitionTime":"2025-11-25T12:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.937593 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.937666 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.937693 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.940730 4715 generic.go:334] "Generic (PLEG): container finished" podID="f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c" containerID="e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4" exitCode=0 Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.940782 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerDied","Data":"e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4"} Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.958963 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.974215 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:41 crc kubenswrapper[4715]: I1125 12:09:41.993084 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:41Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.009460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.009492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.009501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.009515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.009527 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.017568 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.035435 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.050849 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.067331 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.080997 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.093758 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.104109 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.112219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.112433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.112508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.112587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.112710 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.117121 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.133101 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.164921 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.176981 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.215517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.215840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.215908 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.215995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.216144 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.319543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.319598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.319608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.319628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.319641 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.422373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.422408 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.422418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.422436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.422445 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.525364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.525430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.525449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.525475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.525500 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.628752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.628824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.628841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.628863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.628910 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.693575 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.693638 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.693669 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:42 crc kubenswrapper[4715]: E1125 12:09:42.693785 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:42 crc kubenswrapper[4715]: E1125 12:09:42.693883 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:42 crc kubenswrapper[4715]: E1125 12:09:42.694007 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.731993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.732312 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.732376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.732484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.732593 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.835370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.835426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.835439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.835457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.835468 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.938964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.939048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.939059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.939078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.939091 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:42Z","lastTransitionTime":"2025-11-25T12:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.949824 4715 generic.go:334] "Generic (PLEG): container finished" podID="f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c" containerID="2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4" exitCode=0 Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.949909 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerDied","Data":"2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.954641 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} Nov 25 12:09:42 crc kubenswrapper[4715]: I1125 12:09:42.989829 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:42Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.002504 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.016711 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.036152 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.042367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.042414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.042423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.042439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.042452 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.052439 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.066693 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.082836 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.095656 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.110154 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.125758 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.137373 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.145245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.145295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.145305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.145324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.145336 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.147261 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.159051 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.180956 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.248475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.248527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.248539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.248560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.248573 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.351348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.351651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.351662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.351677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.351686 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.453911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.453943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.453954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.453969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.453978 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.557226 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.557283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.557294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.557322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.557336 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.660858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.660927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.660945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.660971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.660984 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.764066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.764118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.764131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.764155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.764175 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.867436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.867511 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.867531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.867559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.867595 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.962528 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerStarted","Data":"4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.970142 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.970224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.970239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.970259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.970270 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:43Z","lastTransitionTime":"2025-11-25T12:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.980096 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:43 crc kubenswrapper[4715]: I1125 12:09:43.994869 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:43Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.010348 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.025430 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.037994 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.053336 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.069742 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.072966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.073047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.073064 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.073087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.073101 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:44Z","lastTransitionTime":"2025-11-25T12:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.083347 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.097865 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.112089 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.127420 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.140942 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.156612 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.176065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.176129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.176145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.176169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.176218 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:44Z","lastTransitionTime":"2025-11-25T12:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.176954 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:44Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.279610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.279669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.279682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.279701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.279715 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:44Z","lastTransitionTime":"2025-11-25T12:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.382616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.382990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.383062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.383142 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.383237 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:44Z","lastTransitionTime":"2025-11-25T12:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.486783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.486848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.486857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.486873 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.486882 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:44Z","lastTransitionTime":"2025-11-25T12:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.589815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.589867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.589883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.589930 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.589947 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:44Z","lastTransitionTime":"2025-11-25T12:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.692348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.692390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.692401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.692419 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.692430 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:44Z","lastTransitionTime":"2025-11-25T12:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.692887 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.692887 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:44 crc kubenswrapper[4715]: E1125 12:09:44.692994 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.693051 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:44 crc kubenswrapper[4715]: E1125 12:09:44.693139 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:44 crc kubenswrapper[4715]: E1125 12:09:44.693339 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.795782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.795833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.795853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.795879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.795920 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:44Z","lastTransitionTime":"2025-11-25T12:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.898636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.898704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.898728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.898760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:44 crc kubenswrapper[4715]: I1125 12:09:44.898783 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:44Z","lastTransitionTime":"2025-11-25T12:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.001740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.001783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.001795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.001815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.001828 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.104413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.104461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.104473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.104497 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.104512 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.207322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.207359 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.207372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.207388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.207399 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.309947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.309987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.309996 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.310014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.310025 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.413414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.413488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.413497 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.413513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.413525 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.516402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.516441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.516451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.516467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.516477 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.619010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.619067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.619080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.619101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.619115 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.722308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.722386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.722416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.722461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.722494 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.826245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.826308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.826319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.826341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.826355 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.929680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.929730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.929746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.929767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.929779 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:45Z","lastTransitionTime":"2025-11-25T12:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.976570 4715 generic.go:334] "Generic (PLEG): container finished" podID="f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c" containerID="4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399" exitCode=0 Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.976653 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerDied","Data":"4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.983842 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} Nov 25 12:09:45 crc kubenswrapper[4715]: I1125 12:09:45.996133 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:45Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.014366 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.030997 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.033512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.033552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.033562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.033577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.033587 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.044722 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.061615 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.075438 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.093293 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.122497 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.141737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.141791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.141813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.141836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.141850 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.145258 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.164920 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.183427 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.213309 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.233854 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.244677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.244746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.244761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.244781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.244798 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.250237 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:46Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.348581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.348673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.348689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.348713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.348731 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.450990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.451040 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.451054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.451076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.451092 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.553889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.553928 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.553937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.553953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.553962 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.656565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.656639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.656654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.656677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.656695 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.693015 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.693090 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.693115 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:46 crc kubenswrapper[4715]: E1125 12:09:46.693247 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:46 crc kubenswrapper[4715]: E1125 12:09:46.693317 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:46 crc kubenswrapper[4715]: E1125 12:09:46.693463 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.759543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.759603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.759617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.759637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.759649 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.862302 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.862341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.862368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.862384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.862395 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.966126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.966168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.966180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.966218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.966237 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:46Z","lastTransitionTime":"2025-11-25T12:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.999692 4715 generic.go:334] "Generic (PLEG): container finished" podID="f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c" containerID="469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2" exitCode=0 Nov 25 12:09:46 crc kubenswrapper[4715]: I1125 12:09:46.999747 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerDied","Data":"469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2"} Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.014516 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.029381 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.045705 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.058293 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.069720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.069787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.069800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.069820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.069834 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:47Z","lastTransitionTime":"2025-11-25T12:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.074970 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.094561 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.109374 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.125931 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.139478 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.152775 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.168319 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.172887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.172942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.172954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.172974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.172986 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:47Z","lastTransitionTime":"2025-11-25T12:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.183423 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.197628 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.219245 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:47Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.278130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.278231 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.278247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.278269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.278284 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:47Z","lastTransitionTime":"2025-11-25T12:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.380969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.381021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.381037 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.381063 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.381081 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:47Z","lastTransitionTime":"2025-11-25T12:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.483826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.483890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.483905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.483931 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.483954 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:47Z","lastTransitionTime":"2025-11-25T12:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.586660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.586708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.586717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.586736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.586747 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:47Z","lastTransitionTime":"2025-11-25T12:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.689462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.689505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.689516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.689531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.689539 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:47Z","lastTransitionTime":"2025-11-25T12:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.694362 4715 scope.go:117] "RemoveContainer" containerID="3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.792916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.792969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.792983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.793001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.793014 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:47Z","lastTransitionTime":"2025-11-25T12:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.896336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.896396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.896412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.896435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:47 crc kubenswrapper[4715]: I1125 12:09:47.896451 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:47Z","lastTransitionTime":"2025-11-25T12:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.000452 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.000513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.000525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.000544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.000557 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.010629 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.010821 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.010847 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.017800 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" event={"ID":"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c","Type":"ContainerStarted","Data":"eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.027472 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.042273 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.060343 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.070795 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.072360 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.076496 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.099258 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.103489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.103527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.103536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.103554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.103565 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.112613 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.125837 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.141468 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.156072 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.169644 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.185269 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.205711 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.206569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.206610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.206623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.206642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.206653 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.221839 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.237168 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.252597 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.272250 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.287111 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.302405 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.309036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.309077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.309091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.309109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.309120 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.317604 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.330405 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.348458 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.367042 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.381509 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.391179 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.391309 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.391358 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:04.391342423 +0000 UTC m=+54.898845444 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.398060 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.411991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.412023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.412031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.412046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.412057 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.414088 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.429666 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.442642 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.455249 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.492391 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.492501 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.492534 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.492558 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.492734 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.492753 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.492769 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.492821 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:04.492803894 +0000 UTC m=+55.000306915 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.493368 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:10:04.493355369 +0000 UTC m=+55.000858390 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.493489 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.493538 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:04.493527193 +0000 UTC m=+55.001030224 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.493613 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.493638 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.493652 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.493682 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:04.493673507 +0000 UTC m=+55.001176528 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.515129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.515170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.515201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.515222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.515236 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.617808 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.617857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.617866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.617885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.617896 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.693743 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.693751 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.693768 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.693899 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.694125 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.694429 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.721032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.721077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.721089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.721105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.721116 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.824852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.824897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.824910 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.824938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.824950 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.834679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.834724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.834736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.834754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.834768 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.849470 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.853462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.853501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.853509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.853524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.853533 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.865538 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.872101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.872173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.872204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.872225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.872241 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.886224 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.890335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.890500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.890578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.890672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.890748 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.906525 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.911492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.911537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.911548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.911566 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.911578 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.926652 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:48Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:48 crc kubenswrapper[4715]: E1125 12:09:48.926832 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.928970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.929014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.929026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.929044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:48 crc kubenswrapper[4715]: I1125 12:09:48.929055 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:48Z","lastTransitionTime":"2025-11-25T12:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.023742 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.026055 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.026805 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.027496 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.031978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.032009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.032019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.032034 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.032045 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.044788 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.064801 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.081061 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.096862 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.113700 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.129116 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.134447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.134479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.134492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.134510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.134523 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.149705 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.163158 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.186535 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.202744 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.217317 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.234015 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.237864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.237911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.237925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.237944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.237956 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.252811 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.276944 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.341147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.341227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.341241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.341261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.341278 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.444482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.444558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.444568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.444587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.444612 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.546911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.546953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.546962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.546977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.546986 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.649890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.649938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.649953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.649972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.649985 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.749466 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f"] Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.750052 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.752273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.752342 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.752360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.752388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.752406 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.753391 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.753625 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.770567 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.784759 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.798371 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.807677 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t9m5\" (UniqueName: \"kubernetes.io/projected/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-kube-api-access-6t9m5\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.807757 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.807815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.807851 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.814878 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.858150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.858212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.858232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.858250 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.858261 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.859347 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.875488 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.891169 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.901593 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.908409 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.908461 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.908527 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t9m5\" (UniqueName: \"kubernetes.io/projected/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-kube-api-access-6t9m5\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.908578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.909469 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.909584 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.912937 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.916694 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.925640 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t9m5\" (UniqueName: \"kubernetes.io/projected/42ea83ae-7f5a-4224-8e39-c2a86950ee7c-kube-api-access-6t9m5\") pod \"ovnkube-control-plane-749d76644c-vwc5f\" (UID: \"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.929607 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.945309 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.959560 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.965323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.965360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.965368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.965383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.965392 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:49Z","lastTransitionTime":"2025-11-25T12:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.975031 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:49 crc kubenswrapper[4715]: I1125 12:09:49.993541 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:49Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.011791 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.029360 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.065381 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.084563 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.084595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.084606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.084624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.084637 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:50Z","lastTransitionTime":"2025-11-25T12:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.187754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.187840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.187855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.187886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.187900 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:50Z","lastTransitionTime":"2025-11-25T12:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.290303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.290650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.290737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.290905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.291000 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:50Z","lastTransitionTime":"2025-11-25T12:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.394885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.394957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.394971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.394995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.395015 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:50Z","lastTransitionTime":"2025-11-25T12:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.496838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.496889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.496903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.496924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.496938 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:50Z","lastTransitionTime":"2025-11-25T12:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.599366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.599411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.599422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.599439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.599450 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:50Z","lastTransitionTime":"2025-11-25T12:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.694149 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.694232 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:50 crc kubenswrapper[4715]: E1125 12:09:50.694369 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:50 crc kubenswrapper[4715]: E1125 12:09:50.694503 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.694783 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:50 crc kubenswrapper[4715]: E1125 12:09:50.695092 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.706168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.706253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.706265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.706317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.706330 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:50Z","lastTransitionTime":"2025-11-25T12:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.724433 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.742726 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.757677 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.774325 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.788640 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.806262 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.811665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.811708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.811721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.811742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.811754 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:50Z","lastTransitionTime":"2025-11-25T12:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.821425 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.838428 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.853904 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.871499 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.886338 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.906041 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.915438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.915495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.915508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.915532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.915546 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:50Z","lastTransitionTime":"2025-11-25T12:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.923955 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.941210 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:50 crc kubenswrapper[4715]: I1125 12:09:50.956793 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:50Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.018503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.018564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.018579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.018599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.018614 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.034689 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" event={"ID":"42ea83ae-7f5a-4224-8e39-c2a86950ee7c","Type":"ContainerStarted","Data":"1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.034743 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" event={"ID":"42ea83ae-7f5a-4224-8e39-c2a86950ee7c","Type":"ContainerStarted","Data":"84ad1d7d6c03bde384bacfa396db093a3358e70a93c768f2137acf64fd1fa4c4"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.121033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.121100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.121112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.121136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.121150 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.224574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.224625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.224635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.224654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.224665 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.249267 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-c8s6d"] Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.249848 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:51 crc kubenswrapper[4715]: E1125 12:09:51.249930 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.271174 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.290169 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.307224 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.324003 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cflzx\" (UniqueName: \"kubernetes.io/projected/3874a0dc-1a53-4587-bb14-7374a9715ae5-kube-api-access-cflzx\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.324051 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.324618 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.327707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.327832 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.327896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.327962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.328035 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.342526 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.355788 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.373637 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.392915 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.406097 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.424879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cflzx\" (UniqueName: \"kubernetes.io/projected/3874a0dc-1a53-4587-bb14-7374a9715ae5-kube-api-access-cflzx\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.424964 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:51 crc kubenswrapper[4715]: E1125 12:09:51.425107 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:51 crc kubenswrapper[4715]: E1125 12:09:51.425177 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs podName:3874a0dc-1a53-4587-bb14-7374a9715ae5 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:51.925155335 +0000 UTC m=+42.432658356 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs") pod "network-metrics-daemon-c8s6d" (UID: "3874a0dc-1a53-4587-bb14-7374a9715ae5") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.430887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.430924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.430936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.430958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.430971 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.453979 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cflzx\" (UniqueName: \"kubernetes.io/projected/3874a0dc-1a53-4587-bb14-7374a9715ae5-kube-api-access-cflzx\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.456370 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.479999 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.494394 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.507262 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.521577 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.534321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.534383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.534396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.534417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.534430 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.537345 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.551535 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:51Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.638046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.638096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.638106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.638124 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.638136 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.741428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.741481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.741492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.741511 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.741522 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.844316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.844371 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.844384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.844403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.844414 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.930847 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:51 crc kubenswrapper[4715]: E1125 12:09:51.931052 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:51 crc kubenswrapper[4715]: E1125 12:09:51.931139 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs podName:3874a0dc-1a53-4587-bb14-7374a9715ae5 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:52.931120574 +0000 UTC m=+43.438623595 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs") pod "network-metrics-daemon-c8s6d" (UID: "3874a0dc-1a53-4587-bb14-7374a9715ae5") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.947424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.947479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.947491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.947538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:51 crc kubenswrapper[4715]: I1125 12:09:51.947551 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:51Z","lastTransitionTime":"2025-11-25T12:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.040677 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/0.log" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.045051 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257" exitCode=1 Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.045111 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.046641 4715 scope.go:117] "RemoveContainer" containerID="623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.047579 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" event={"ID":"42ea83ae-7f5a-4224-8e39-c2a86950ee7c","Type":"ContainerStarted","Data":"c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.049711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.049762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.049776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.049795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.049807 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.063973 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.078487 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.092736 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.109465 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.124146 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.137570 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.152779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.152818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.152830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.152847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.152858 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.153577 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.164299 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.176872 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.189544 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.205020 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.216294 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.233153 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.257317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.257360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.257374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.257392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.257406 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.258445 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI1125 12:09:50.926405 6009 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.926529 6009 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 12:09:50.926833 6009 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927025 6009 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927380 6009 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:50.927402 6009 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:50.927431 6009 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 12:09:50.927435 6009 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:50.927445 6009 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 12:09:50.927462 6009 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:50.927468 6009 factory.go:656] Stopping watch factory\\\\nI1125 12:09:50.927482 6009 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 12:09:50.927495 6009 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.272790 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.289517 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.303003 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.317020 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.332242 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.358583 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI1125 12:09:50.926405 6009 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.926529 6009 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 12:09:50.926833 6009 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927025 6009 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927380 6009 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:50.927402 6009 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:50.927431 6009 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 12:09:50.927435 6009 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:50.927445 6009 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 12:09:50.927462 6009 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:50.927468 6009 factory.go:656] Stopping watch factory\\\\nI1125 12:09:50.927482 6009 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 12:09:50.927495 6009 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.359805 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.359840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.359849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.359864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.359875 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.373473 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.393246 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.408778 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.423449 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.446580 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.462600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.462653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.462663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.462679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.462690 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.464245 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.483023 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.505851 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.520886 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.537364 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.549015 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.565643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.565694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.565706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.565725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.565737 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.567141 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:52Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.668563 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.668633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.668651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.668678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.668694 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.692950 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.693083 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:52 crc kubenswrapper[4715]: E1125 12:09:52.693140 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.693178 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.693341 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:52 crc kubenswrapper[4715]: E1125 12:09:52.693372 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:09:52 crc kubenswrapper[4715]: E1125 12:09:52.693430 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:52 crc kubenswrapper[4715]: E1125 12:09:52.693464 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.772114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.772154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.772162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.772203 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.772217 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.876404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.876458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.876468 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.876486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.876496 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.942513 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:52 crc kubenswrapper[4715]: E1125 12:09:52.942719 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:52 crc kubenswrapper[4715]: E1125 12:09:52.942803 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs podName:3874a0dc-1a53-4587-bb14-7374a9715ae5 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:54.942782624 +0000 UTC m=+45.450285645 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs") pod "network-metrics-daemon-c8s6d" (UID: "3874a0dc-1a53-4587-bb14-7374a9715ae5") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.979585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.979647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.979657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.979675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:52 crc kubenswrapper[4715]: I1125 12:09:52.979686 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:52Z","lastTransitionTime":"2025-11-25T12:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.053403 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/1.log" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.053887 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/0.log" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.057271 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829" exitCode=1 Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.057361 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.057440 4715 scope.go:117] "RemoveContainer" containerID="623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.058515 4715 scope.go:117] "RemoveContainer" containerID="109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829" Nov 25 12:09:53 crc kubenswrapper[4715]: E1125 12:09:53.058713 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.075819 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.082487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.082549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.082563 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.082589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.082627 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:53Z","lastTransitionTime":"2025-11-25T12:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.088036 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.103240 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.118938 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.134942 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.148501 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.167970 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.186247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.186293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.186309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.186329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.186340 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:53Z","lastTransitionTime":"2025-11-25T12:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.188507 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI1125 12:09:50.926405 6009 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.926529 6009 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 12:09:50.926833 6009 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927025 6009 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927380 6009 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:50.927402 6009 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:50.927431 6009 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 12:09:50.927435 6009 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:50.927445 6009 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 12:09:50.927462 6009 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:50.927468 6009 factory.go:656] Stopping watch factory\\\\nI1125 12:09:50.927482 6009 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 12:09:50.927495 6009 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"message\\\":\\\"ndler 9 for removal\\\\nI1125 12:09:52.898257 6224 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 12:09:52.898304 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 12:09:52.898258 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 12:09:52.898331 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 12:09:52.898366 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:52.898377 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:52.898387 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 12:09:52.898407 6224 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 12:09:52.898413 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 12:09:52.898415 6224 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:52.898439 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:52.898457 6224 factory.go:656] Stopping watch factory\\\\nI1125 12:09:52.898469 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 12:09:52.898507 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1125 12:09:52.898558 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 12:09:52.898639 6224 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.200400 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.214479 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.229671 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.243031 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.257973 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.271026 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.283990 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.289143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.289215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.289225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.289245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.289257 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:53Z","lastTransitionTime":"2025-11-25T12:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.300992 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:53Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.392544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.392591 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.392600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.392615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.392627 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:53Z","lastTransitionTime":"2025-11-25T12:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.496250 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.496328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.496354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.496387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.496411 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:53Z","lastTransitionTime":"2025-11-25T12:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.602726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.602781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.602794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.602812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.602822 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:53Z","lastTransitionTime":"2025-11-25T12:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.705892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.705984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.705995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.706013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.706024 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:53Z","lastTransitionTime":"2025-11-25T12:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.809304 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.809366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.809381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.809402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.809419 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:53Z","lastTransitionTime":"2025-11-25T12:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.912228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.912284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.912294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.912340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:53 crc kubenswrapper[4715]: I1125 12:09:53.912353 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:53Z","lastTransitionTime":"2025-11-25T12:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.015726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.015780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.015795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.015814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.015828 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.063173 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/1.log" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.119662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.119859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.119893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.119927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.119949 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.223146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.223205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.223218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.223237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.223250 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.325423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.325470 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.325485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.325504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.325516 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.428787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.428826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.428835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.428850 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.428860 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.531073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.531149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.531162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.531202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.531217 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.634060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.634134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.634143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.634162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.634173 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.693319 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:54 crc kubenswrapper[4715]: E1125 12:09:54.693508 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.693575 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.693586 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:54 crc kubenswrapper[4715]: E1125 12:09:54.693781 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.693599 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:54 crc kubenswrapper[4715]: E1125 12:09:54.693981 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:54 crc kubenswrapper[4715]: E1125 12:09:54.694079 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.737840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.737891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.737903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.737958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.737979 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.840939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.840993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.841005 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.841023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.841032 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.944046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.944098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.944110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.944136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.944148 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:54Z","lastTransitionTime":"2025-11-25T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:54 crc kubenswrapper[4715]: I1125 12:09:54.967063 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:54 crc kubenswrapper[4715]: E1125 12:09:54.967301 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:54 crc kubenswrapper[4715]: E1125 12:09:54.967411 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs podName:3874a0dc-1a53-4587-bb14-7374a9715ae5 nodeName:}" failed. No retries permitted until 2025-11-25 12:09:58.967387198 +0000 UTC m=+49.474890239 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs") pod "network-metrics-daemon-c8s6d" (UID: "3874a0dc-1a53-4587-bb14-7374a9715ae5") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.047030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.047093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.047109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.047132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.047149 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.156216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.156446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.156459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.156478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.156492 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.261234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.261307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.261323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.261349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.261389 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.365132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.365239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.365262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.365291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.365311 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.468663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.468749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.468761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.468787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.468801 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.571400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.571738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.571846 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.571956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.572045 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.674462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.674503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.674517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.674534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.674546 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.776882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.776929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.776941 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.776962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.776973 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.879535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.879576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.879585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.879599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.879609 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.981616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.981666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.981684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.981704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:55 crc kubenswrapper[4715]: I1125 12:09:55.981716 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:55Z","lastTransitionTime":"2025-11-25T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.084227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.084326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.084353 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.084377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.084393 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:56Z","lastTransitionTime":"2025-11-25T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.187675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.187738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.187747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.187766 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.187775 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:56Z","lastTransitionTime":"2025-11-25T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.290231 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.290286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.290302 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.290323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.290338 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:56Z","lastTransitionTime":"2025-11-25T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.393338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.393386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.393395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.393413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.393422 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:56Z","lastTransitionTime":"2025-11-25T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.495053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.495093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.495105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.495122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.495133 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:56Z","lastTransitionTime":"2025-11-25T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.597599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.597671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.597685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.597705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.597719 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:56Z","lastTransitionTime":"2025-11-25T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.693457 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.693648 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.693486 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:56 crc kubenswrapper[4715]: E1125 12:09:56.693734 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.693772 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:56 crc kubenswrapper[4715]: E1125 12:09:56.694004 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:56 crc kubenswrapper[4715]: E1125 12:09:56.694146 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:56 crc kubenswrapper[4715]: E1125 12:09:56.694399 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.700227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.700326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.700344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.700373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.700392 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:56Z","lastTransitionTime":"2025-11-25T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.803128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.803170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.803180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.803216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.803236 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:56Z","lastTransitionTime":"2025-11-25T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.905671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.905722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.905745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.905772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:56 crc kubenswrapper[4715]: I1125 12:09:56.905784 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:56Z","lastTransitionTime":"2025-11-25T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.008147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.008226 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.008237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.008255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.008269 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.110635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.110690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.110702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.110724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.110735 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.214334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.214378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.214392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.214409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.214419 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.317324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.317358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.317367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.317384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.317395 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.420600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.420677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.420686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.420705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.420715 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.524164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.524241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.524252 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.524271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.524286 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.627054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.627129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.627147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.627174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.627247 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.730009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.730090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.730109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.730136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.730155 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.834002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.834064 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.834078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.834097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.834108 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.937050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.937099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.937117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.937137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:57 crc kubenswrapper[4715]: I1125 12:09:57.937148 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:57Z","lastTransitionTime":"2025-11-25T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.040290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.040351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.040365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.040385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.040397 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.142999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.143055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.143066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.143087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.143101 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.245888 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.245932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.245943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.245960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.245972 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.348108 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.348153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.348167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.348202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.348213 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.450922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.450968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.450978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.450995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.451006 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.553671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.553711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.553729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.553752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.553770 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.656378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.656453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.656467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.656493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.656506 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.693663 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.693699 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:09:58 crc kubenswrapper[4715]: E1125 12:09:58.693883 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.693721 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:09:58 crc kubenswrapper[4715]: E1125 12:09:58.693972 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.693699 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:58 crc kubenswrapper[4715]: E1125 12:09:58.694008 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:09:58 crc kubenswrapper[4715]: E1125 12:09:58.694057 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.758789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.758850 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.758862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.758879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.758890 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.861175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.861251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.861264 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.861283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.861296 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.964444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.964501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.964524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.964555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:58 crc kubenswrapper[4715]: I1125 12:09:58.964568 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:58Z","lastTransitionTime":"2025-11-25T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.009992 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:09:59 crc kubenswrapper[4715]: E1125 12:09:59.010286 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:59 crc kubenswrapper[4715]: E1125 12:09:59.010435 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs podName:3874a0dc-1a53-4587-bb14-7374a9715ae5 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:07.010396526 +0000 UTC m=+57.517899587 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs") pod "network-metrics-daemon-c8s6d" (UID: "3874a0dc-1a53-4587-bb14-7374a9715ae5") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.067260 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.067322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.067334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.067357 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.067371 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.106469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.106519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.106533 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.106553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.106565 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: E1125 12:09:59.120917 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:59Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.124834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.124878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.124891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.124911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.124923 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: E1125 12:09:59.135341 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:59Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.138572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.138609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.138622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.138638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.138650 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: E1125 12:09:59.152419 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:59Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.156261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.156319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.156330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.156364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.156374 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: E1125 12:09:59.169372 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:59Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.176611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.176663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.176680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.176709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.176727 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: E1125 12:09:59.192285 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:09:59Z is after 2025-08-24T17:21:41Z" Nov 25 12:09:59 crc kubenswrapper[4715]: E1125 12:09:59.192415 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.194298 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.194348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.194361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.194383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.194396 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.297269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.297337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.297349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.297370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.297384 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.400281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.400333 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.400344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.400363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.400376 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.503015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.503066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.503086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.503108 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.503121 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.605515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.605561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.605571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.605586 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.605594 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.707958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.708016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.708032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.708053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.708071 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.812070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.812208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.812235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.812269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.812293 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.915608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.915661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.915691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.915739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:09:59 crc kubenswrapper[4715]: I1125 12:09:59.915755 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:09:59Z","lastTransitionTime":"2025-11-25T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.018598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.018655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.018677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.018703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.018719 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.121360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.121409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.121418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.121436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.121448 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.223488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.223534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.223545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.223561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.223571 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.326271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.326324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.326336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.326358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.326377 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.428872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.428919 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.428934 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.428956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.428967 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.531678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.531739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.531758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.531780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.531794 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.634510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.634854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.634950 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.635038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.635119 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.693547 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.693697 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:00 crc kubenswrapper[4715]: E1125 12:10:00.693874 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.693950 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.693953 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:00 crc kubenswrapper[4715]: E1125 12:10:00.694152 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:00 crc kubenswrapper[4715]: E1125 12:10:00.694267 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:00 crc kubenswrapper[4715]: E1125 12:10:00.694121 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.711628 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.723997 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.736784 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.737593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.737628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.737638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.737652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.737662 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.748524 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.760108 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.770878 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.782042 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.799051 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI1125 12:09:50.926405 6009 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.926529 6009 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 12:09:50.926833 6009 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927025 6009 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927380 6009 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:50.927402 6009 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:50.927431 6009 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 12:09:50.927435 6009 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:50.927445 6009 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 12:09:50.927462 6009 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:50.927468 6009 factory.go:656] Stopping watch factory\\\\nI1125 12:09:50.927482 6009 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 12:09:50.927495 6009 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"message\\\":\\\"ndler 9 for removal\\\\nI1125 12:09:52.898257 6224 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 12:09:52.898304 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 12:09:52.898258 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 12:09:52.898331 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 12:09:52.898366 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:52.898377 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:52.898387 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 12:09:52.898407 6224 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 12:09:52.898413 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 12:09:52.898415 6224 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:52.898439 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:52.898457 6224 factory.go:656] Stopping watch factory\\\\nI1125 12:09:52.898469 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 12:09:52.898507 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1125 12:09:52.898558 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 12:09:52.898639 6224 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.809901 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.823651 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.836541 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.840965 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.841026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.841038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.841058 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.841072 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.850631 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.861342 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.874965 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.886725 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.897375 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:00Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.943748 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.943806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.943820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.943843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:00 crc kubenswrapper[4715]: I1125 12:10:00.943855 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:00Z","lastTransitionTime":"2025-11-25T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.046362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.046407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.046416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.046432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.046442 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.149457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.149510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.149519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.149537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.149546 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.251883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.251916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.251925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.251938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.251947 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.354575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.354626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.354642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.354664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.354677 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.457229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.457276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.457288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.457307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.457318 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.560605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.560671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.560694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.560723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.560740 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.663372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.663445 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.663464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.663489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.663508 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.766851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.766909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.766918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.766970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.766981 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.870061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.870159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.870198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.870218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.870231 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.973043 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.973099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.973113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.973134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:01 crc kubenswrapper[4715]: I1125 12:10:01.973146 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:01Z","lastTransitionTime":"2025-11-25T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.075902 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.075954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.075963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.075980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.075990 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:02Z","lastTransitionTime":"2025-11-25T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.179673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.179721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.179735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.179755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.179770 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:02Z","lastTransitionTime":"2025-11-25T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.254300 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.268210 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.279398 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.282230 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.282289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.282305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.282327 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.282337 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:02Z","lastTransitionTime":"2025-11-25T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.296134 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.317376 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://623b3c3f260f674640ae1f0da5742721abd2f1af26fa3a4f65a7abe98e5ba257\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"factory.go:160\\\\nI1125 12:09:50.926405 6009 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.926529 6009 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 12:09:50.926833 6009 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927025 6009 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 12:09:50.927380 6009 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:50.927402 6009 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:50.927431 6009 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 12:09:50.927435 6009 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:50.927445 6009 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 12:09:50.927462 6009 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:50.927468 6009 factory.go:656] Stopping watch factory\\\\nI1125 12:09:50.927482 6009 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 12:09:50.927495 6009 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"message\\\":\\\"ndler 9 for removal\\\\nI1125 12:09:52.898257 6224 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 12:09:52.898304 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 12:09:52.898258 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 12:09:52.898331 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 12:09:52.898366 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:52.898377 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:52.898387 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 12:09:52.898407 6224 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 12:09:52.898413 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 12:09:52.898415 6224 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:52.898439 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:52.898457 6224 factory.go:656] Stopping watch factory\\\\nI1125 12:09:52.898469 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 12:09:52.898507 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1125 12:09:52.898558 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 12:09:52.898639 6224 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.329791 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.342872 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.358214 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.371537 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.384048 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.385032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.385062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.385072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.385089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.385098 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:02Z","lastTransitionTime":"2025-11-25T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.398573 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.412601 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.428585 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.446099 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.460582 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.471911 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.482026 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:02Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.488216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.488255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.488266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.488282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.488292 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:02Z","lastTransitionTime":"2025-11-25T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.591741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.591784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.591794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.591811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.591827 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:02Z","lastTransitionTime":"2025-11-25T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.693044 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.693122 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.693166 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:02 crc kubenswrapper[4715]: E1125 12:10:02.693287 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.693307 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:02 crc kubenswrapper[4715]: E1125 12:10:02.693412 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:02 crc kubenswrapper[4715]: E1125 12:10:02.693514 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:02 crc kubenswrapper[4715]: E1125 12:10:02.693602 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.694806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.694852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.694867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.694886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.694899 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:02Z","lastTransitionTime":"2025-11-25T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.797556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.797905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.798023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.798152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.798307 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:02Z","lastTransitionTime":"2025-11-25T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.902621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.902662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.902671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.902688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:02 crc kubenswrapper[4715]: I1125 12:10:02.902701 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:02Z","lastTransitionTime":"2025-11-25T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.005850 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.005918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.005929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.005951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.005963 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.108211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.108267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.108285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.108307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.108318 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.211679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.211771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.211780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.211797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.211810 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.314465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.314518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.314528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.314545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.314556 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.418002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.418048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.418059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.418078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.418091 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.520798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.520843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.520852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.520868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.520879 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.624030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.624088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.624101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.624120 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.624133 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.710679 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.711627 4715 scope.go:117] "RemoveContainer" containerID="109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.724582 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.726577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.726610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.726622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.726642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.726655 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.743047 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.764697 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"message\\\":\\\"ndler 9 for removal\\\\nI1125 12:09:52.898257 6224 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 12:09:52.898304 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 12:09:52.898258 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 12:09:52.898331 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 12:09:52.898366 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:52.898377 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:52.898387 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 12:09:52.898407 6224 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 12:09:52.898413 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 12:09:52.898415 6224 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:52.898439 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:52.898457 6224 factory.go:656] Stopping watch factory\\\\nI1125 12:09:52.898469 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 12:09:52.898507 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1125 12:09:52.898558 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 12:09:52.898639 6224 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.781512 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.801736 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.814228 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.829551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.829616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.829629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.829650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.829661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.840162 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.869429 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.884111 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.901178 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.916555 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.932893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.932935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.932945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.932964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.932977 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:03Z","lastTransitionTime":"2025-11-25T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.938230 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.951023 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.964564 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.978287 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:03 crc kubenswrapper[4715]: I1125 12:10:03.991490 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:03Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.035505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.035548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.035559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.035577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.035588 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.110934 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/1.log" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.120054 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.120632 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.138323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.138372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.138385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.138403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.138415 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.139057 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.156468 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.170059 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.188669 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.202380 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.215411 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.225489 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.236410 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.241425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.241481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.241491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.241510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.241522 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.255824 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"message\\\":\\\"ndler 9 for removal\\\\nI1125 12:09:52.898257 6224 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 12:09:52.898304 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 12:09:52.898258 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 12:09:52.898331 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 12:09:52.898366 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:52.898377 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:52.898387 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 12:09:52.898407 6224 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 12:09:52.898413 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 12:09:52.898415 6224 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:52.898439 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:52.898457 6224 factory.go:656] Stopping watch factory\\\\nI1125 12:09:52.898469 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 12:09:52.898507 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1125 12:09:52.898558 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 12:09:52.898639 6224 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.271569 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.300763 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.316127 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.330649 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.343316 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.344483 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.344538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.344553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.344574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.344586 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.356539 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.375318 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:04Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.447871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.447925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.447934 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.447951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.447963 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.469619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.469753 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.469830 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:36.469807529 +0000 UTC m=+86.977310560 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.550876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.550933 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.550945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.550964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.550975 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.570285 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.570429 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570521 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:10:36.57048304 +0000 UTC m=+87.077986051 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570535 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.570572 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.570613 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570668 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:36.570643774 +0000 UTC m=+87.078146795 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570746 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570771 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570785 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570786 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570812 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570826 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570849 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:36.570837069 +0000 UTC m=+87.078340090 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.570869 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:36.570862329 +0000 UTC m=+87.078365450 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.655004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.655055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.655066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.655085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.655099 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.693948 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.694014 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.694096 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.694136 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.694130 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.694288 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.694504 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:04 crc kubenswrapper[4715]: E1125 12:10:04.694614 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.757856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.757906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.757914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.757930 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.757939 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.860536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.860602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.860614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.860631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.860643 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.963314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.963366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.963378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.963399 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:04 crc kubenswrapper[4715]: I1125 12:10:04.963413 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:04Z","lastTransitionTime":"2025-11-25T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.066175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.066270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.066288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.066313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.066330 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.125696 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/2.log" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.126430 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/1.log" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.129486 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386" exitCode=1 Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.129537 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.129582 4715 scope.go:117] "RemoveContainer" containerID="109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.130454 4715 scope.go:117] "RemoveContainer" containerID="d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386" Nov 25 12:10:05 crc kubenswrapper[4715]: E1125 12:10:05.130706 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.148024 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.162311 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.169646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.169686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.169697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.169715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.169727 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.175893 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.186934 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.200858 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.211947 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.223910 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.238064 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.250983 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.262835 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.272970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.273024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.273039 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.273036 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.273260 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.273276 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.285091 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.305447 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"message\\\":\\\"ndler 9 for removal\\\\nI1125 12:09:52.898257 6224 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 12:09:52.898304 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 12:09:52.898258 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 12:09:52.898331 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 12:09:52.898366 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:52.898377 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:52.898387 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 12:09:52.898407 6224 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 12:09:52.898413 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 12:09:52.898415 6224 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:52.898439 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:52.898457 6224 factory.go:656] Stopping watch factory\\\\nI1125 12:09:52.898469 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 12:09:52.898507 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1125 12:09:52.898558 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 12:09:52.898639 6224 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.316621 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.333258 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.346253 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.379701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.379981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.380098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.380157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.380209 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.483007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.483061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.483071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.483091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.483105 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.539813 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.550545 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.551542 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.564683 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.575775 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.586406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.586477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.586495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.586524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.586541 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.591247 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.606225 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.621432 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.632665 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.646528 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.665766 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://109c93656f6519cc51f66b4e7be81fe3b8e0cdb043cf9ebebca8648f8329a829\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"message\\\":\\\"ndler 9 for removal\\\\nI1125 12:09:52.898257 6224 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 12:09:52.898304 6224 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 12:09:52.898258 6224 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 12:09:52.898331 6224 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 12:09:52.898366 6224 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 12:09:52.898377 6224 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 12:09:52.898387 6224 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 12:09:52.898407 6224 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 12:09:52.898413 6224 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 12:09:52.898415 6224 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 12:09:52.898439 6224 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 12:09:52.898457 6224 factory.go:656] Stopping watch factory\\\\nI1125 12:09:52.898469 6224 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 12:09:52.898507 6224 ovnkube.go:599] Stopped ovnkube\\\\nI1125 12:09:52.898558 6224 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 12:09:52.898639 6224 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.680068 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.689079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.689119 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.689132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.689151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.689163 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.692418 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.705428 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.718329 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.728165 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.738294 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.752955 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:05Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.792887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.792940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.792949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.792971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.792983 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.896427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.896482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.896494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.896516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.896528 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.999609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.999644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.999652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.999670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:05 crc kubenswrapper[4715]: I1125 12:10:05.999678 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:05Z","lastTransitionTime":"2025-11-25T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.102865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.102942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.102957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.102997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.103015 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:06Z","lastTransitionTime":"2025-11-25T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.136126 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/2.log" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.140367 4715 scope.go:117] "RemoveContainer" containerID="d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386" Nov 25 12:10:06 crc kubenswrapper[4715]: E1125 12:10:06.140646 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.153620 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.167994 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.182022 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.199092 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.214793 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.217804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.217849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.217859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.217876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.217887 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:06Z","lastTransitionTime":"2025-11-25T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.229507 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.240506 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.253912 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.272442 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.285847 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.300800 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.319084 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.321758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.321798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.321811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.321830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.321847 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:06Z","lastTransitionTime":"2025-11-25T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.335203 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.348492 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.361466 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.374297 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.388544 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:06Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.424709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.424758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.424772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.424792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.424804 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:06Z","lastTransitionTime":"2025-11-25T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.528072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.528126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.528141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.528166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.528215 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:06Z","lastTransitionTime":"2025-11-25T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.631798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.631848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.631856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.631874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.631884 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:06Z","lastTransitionTime":"2025-11-25T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.693162 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.693298 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.693299 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:06 crc kubenswrapper[4715]: E1125 12:10:06.693479 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.693630 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:06 crc kubenswrapper[4715]: E1125 12:10:06.693774 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:06 crc kubenswrapper[4715]: E1125 12:10:06.693817 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:06 crc kubenswrapper[4715]: E1125 12:10:06.693904 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.734745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.734905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.734922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.734949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.734963 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:06Z","lastTransitionTime":"2025-11-25T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.837319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.837382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.837398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.837422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.837438 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:06Z","lastTransitionTime":"2025-11-25T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.940665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.940712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.940723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.940746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:06 crc kubenswrapper[4715]: I1125 12:10:06.940758 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:06Z","lastTransitionTime":"2025-11-25T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.044056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.044119 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.044132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.044154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.044167 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.096751 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:07 crc kubenswrapper[4715]: E1125 12:10:07.096956 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:10:07 crc kubenswrapper[4715]: E1125 12:10:07.097037 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs podName:3874a0dc-1a53-4587-bb14-7374a9715ae5 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:23.097011668 +0000 UTC m=+73.604514689 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs") pod "network-metrics-daemon-c8s6d" (UID: "3874a0dc-1a53-4587-bb14-7374a9715ae5") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.147479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.147524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.147534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.147549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.147560 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.250633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.250719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.250739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.250768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.250786 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.353658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.353708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.353720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.353737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.353746 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.457220 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.457274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.457285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.457307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.457318 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.560389 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.560465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.560489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.560527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.560556 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.663220 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.663267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.663278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.663294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.663304 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.766487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.766562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.766587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.766611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.766624 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.869537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.869593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.869605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.869628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.869642 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.972968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.973023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.973035 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.973055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:07 crc kubenswrapper[4715]: I1125 12:10:07.973071 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:07Z","lastTransitionTime":"2025-11-25T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.076167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.076235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.076246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.076268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.076282 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:08Z","lastTransitionTime":"2025-11-25T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.179336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.179502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.179525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.179579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.179593 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:08Z","lastTransitionTime":"2025-11-25T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.282694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.282772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.282792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.282818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.282837 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:08Z","lastTransitionTime":"2025-11-25T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.385603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.385657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.385672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.385696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.385710 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:08Z","lastTransitionTime":"2025-11-25T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.488310 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.488362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.488375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.488394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.488407 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:08Z","lastTransitionTime":"2025-11-25T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.591377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.591488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.591510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.591544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.591568 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:08Z","lastTransitionTime":"2025-11-25T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.693917 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.693991 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.694003 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:08 crc kubenswrapper[4715]: E1125 12:10:08.694089 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:08 crc kubenswrapper[4715]: E1125 12:10:08.694263 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:08 crc kubenswrapper[4715]: E1125 12:10:08.694480 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.694667 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:08 crc kubenswrapper[4715]: E1125 12:10:08.694839 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.695489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.695544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.695557 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.695578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.695596 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:08Z","lastTransitionTime":"2025-11-25T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.799172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.799272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.799284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.799309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.799322 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:08Z","lastTransitionTime":"2025-11-25T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.901937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.901984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.901993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.902010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:08 crc kubenswrapper[4715]: I1125 12:10:08.902021 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:08Z","lastTransitionTime":"2025-11-25T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.005545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.005609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.005618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.005635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.005644 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.109085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.109152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.109170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.109228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.109245 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.212323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.212409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.212433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.212463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.212484 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.234604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.234642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.234653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.234671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.234683 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: E1125 12:10:09.247096 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:09Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.251781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.251828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.251839 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.251859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.251870 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: E1125 12:10:09.267682 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:09Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.272745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.272808 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.272825 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.272847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.272863 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: E1125 12:10:09.287302 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:09Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.291739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.291799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.291812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.291834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.291848 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: E1125 12:10:09.304245 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:09Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.307961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.308020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.308033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.308055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.308069 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: E1125 12:10:09.320126 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:09Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:09 crc kubenswrapper[4715]: E1125 12:10:09.320267 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.322325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.322346 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.322355 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.322370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.322381 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.424645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.424707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.424718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.424737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.424752 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.527783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.527835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.527844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.527861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.527874 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.630705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.630761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.630773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.630818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.630834 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.733892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.733936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.733948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.733967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.733980 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.837204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.837255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.837273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.837293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.837305 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.940635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.940713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.940740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.940768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:09 crc kubenswrapper[4715]: I1125 12:10:09.940786 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:09Z","lastTransitionTime":"2025-11-25T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.043487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.043570 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.043596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.043623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.043642 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.147020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.147081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.147090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.147108 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.147118 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.250628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.250687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.250701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.250720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.250733 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.353351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.353390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.353399 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.353415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.353424 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.456600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.456642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.456654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.456673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.456684 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.559335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.559370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.559379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.559394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.559404 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.662506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.662556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.662569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.662585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.662596 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.692890 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.692916 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.692952 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.693060 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:10 crc kubenswrapper[4715]: E1125 12:10:10.693161 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:10 crc kubenswrapper[4715]: E1125 12:10:10.693238 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:10 crc kubenswrapper[4715]: E1125 12:10:10.693307 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:10 crc kubenswrapper[4715]: E1125 12:10:10.693462 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.705337 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.717218 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.726506 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.739464 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.750801 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.762684 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.765474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.765541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.765552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.765597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.765608 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.777577 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.793414 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.816131 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.830260 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.842690 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.853406 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.866287 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.868708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.868764 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.868774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.868790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.868799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.879628 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.893231 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.905268 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.919978 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:10Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.970978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.971015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.971023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.971039 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:10 crc kubenswrapper[4715]: I1125 12:10:10.971049 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:10Z","lastTransitionTime":"2025-11-25T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.073777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.073826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.073849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.073873 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.073888 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:11Z","lastTransitionTime":"2025-11-25T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.176343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.176649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.176710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.176788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.176875 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:11Z","lastTransitionTime":"2025-11-25T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.280144 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.280609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.280767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.280977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.281219 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:11Z","lastTransitionTime":"2025-11-25T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.384158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.384274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.384293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.384323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.384341 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:11Z","lastTransitionTime":"2025-11-25T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.486929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.486988 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.487001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.487022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.487037 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:11Z","lastTransitionTime":"2025-11-25T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.590254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.590305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.590315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.590332 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.590345 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:11Z","lastTransitionTime":"2025-11-25T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.693214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.693503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.693512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.693527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.693539 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:11Z","lastTransitionTime":"2025-11-25T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.796286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.796336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.796350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.796369 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.796384 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:11Z","lastTransitionTime":"2025-11-25T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.898841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.898904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.898914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.898932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:11 crc kubenswrapper[4715]: I1125 12:10:11.898941 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:11Z","lastTransitionTime":"2025-11-25T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.002085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.002158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.002171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.002221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.002237 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.104895 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.104947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.104959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.104980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.105005 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.208684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.208751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.208765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.208785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.208797 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.312012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.312100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.312119 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.312146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.312165 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.416934 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.416997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.417012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.417035 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.417052 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.519864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.519901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.519909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.519924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.519933 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.622344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.622408 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.622423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.622444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.622458 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.693304 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.693367 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.693304 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.693411 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:12 crc kubenswrapper[4715]: E1125 12:10:12.693521 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:12 crc kubenswrapper[4715]: E1125 12:10:12.693619 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:12 crc kubenswrapper[4715]: E1125 12:10:12.693671 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:12 crc kubenswrapper[4715]: E1125 12:10:12.693689 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.725102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.725155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.725165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.725197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.725209 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.827799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.827861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.827874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.827891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.827908 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.931631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.931698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.931711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.931739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:12 crc kubenswrapper[4715]: I1125 12:10:12.931753 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:12Z","lastTransitionTime":"2025-11-25T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.035047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.035124 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.035139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.035163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.035200 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.139172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.139259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.139273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.139294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.139308 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.242319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.242390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.242406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.242433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.242448 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.345875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.345941 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.345958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.345980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.345995 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.449353 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.449423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.449436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.449460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.449476 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.552380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.552435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.552450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.552474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.552491 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.655093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.655224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.655243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.655269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.655287 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.757772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.757823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.757834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.757854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.757865 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.860774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.860826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.860839 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.860858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.860870 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.963616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.963675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.963695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.963713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:13 crc kubenswrapper[4715]: I1125 12:10:13.963723 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:13Z","lastTransitionTime":"2025-11-25T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.067071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.067130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.067140 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.067158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.067168 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.170143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.170221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.170241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.170262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.170276 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.273145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.273212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.273223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.273243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.273261 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.375101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.375155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.375163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.375178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.375201 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.477671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.477719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.477731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.477747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.477758 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.582269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.582355 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.582371 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.582391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.582410 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.685889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.685932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.685942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.685959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.685969 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.695827 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:14 crc kubenswrapper[4715]: E1125 12:10:14.695948 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.696093 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:14 crc kubenswrapper[4715]: E1125 12:10:14.696138 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.696256 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:14 crc kubenswrapper[4715]: E1125 12:10:14.696295 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.696385 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:14 crc kubenswrapper[4715]: E1125 12:10:14.696431 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.787941 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.787972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.787980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.787994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.788004 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.891725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.891778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.891792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.891812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.891825 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.995279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.995340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.995351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.995385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:14 crc kubenswrapper[4715]: I1125 12:10:14.995396 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:14Z","lastTransitionTime":"2025-11-25T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.099169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.099254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.099283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.099303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.099318 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:15Z","lastTransitionTime":"2025-11-25T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.202942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.203011 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.203026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.203043 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.203055 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:15Z","lastTransitionTime":"2025-11-25T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.306165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.306255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.306268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.306288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.306302 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:15Z","lastTransitionTime":"2025-11-25T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.408853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.408896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.408907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.408925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.408965 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:15Z","lastTransitionTime":"2025-11-25T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.511328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.511379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.511390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.511405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.511414 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:15Z","lastTransitionTime":"2025-11-25T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.615067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.615124 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.615138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.615159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.615172 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:15Z","lastTransitionTime":"2025-11-25T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.717416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.717723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.717818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.717913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.718009 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:15Z","lastTransitionTime":"2025-11-25T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.821875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.821938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.821953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.821975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.821991 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:15Z","lastTransitionTime":"2025-11-25T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.925591 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.925932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.926032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.926102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:15 crc kubenswrapper[4715]: I1125 12:10:15.926164 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:15Z","lastTransitionTime":"2025-11-25T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.040634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.040680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.040689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.040707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.040717 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.143426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.143480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.143496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.143518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.143533 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.246367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.246416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.246425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.246444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.246455 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.349474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.349553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.349578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.349610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.349631 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.452271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.452326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.452339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.452358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.452369 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.554728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.554768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.554779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.554794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.554805 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.658168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.658248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.658259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.658284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.658298 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.693895 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.693958 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.693981 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.694054 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:16 crc kubenswrapper[4715]: E1125 12:10:16.694110 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:16 crc kubenswrapper[4715]: E1125 12:10:16.694231 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:16 crc kubenswrapper[4715]: E1125 12:10:16.694315 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:16 crc kubenswrapper[4715]: E1125 12:10:16.694336 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.760897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.761266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.761344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.761442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.761534 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.863647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.864401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.864526 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.864626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.864716 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.967654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.967703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.967715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.967735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:16 crc kubenswrapper[4715]: I1125 12:10:16.967748 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:16Z","lastTransitionTime":"2025-11-25T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.073596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.073652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.073666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.073686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.073703 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.177212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.177287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.177306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.177335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.177352 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.280900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.280946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.280957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.280976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.280989 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.382616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.382653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.382663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.382678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.382686 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.484976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.485021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.485030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.485047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.485059 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.588221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.588265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.588278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.588296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.588313 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.691138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.691210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.691225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.691244 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.691254 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.694380 4715 scope.go:117] "RemoveContainer" containerID="d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386" Nov 25 12:10:17 crc kubenswrapper[4715]: E1125 12:10:17.694581 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.793690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.793733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.793743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.793758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.793767 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.896172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.896233 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.896245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.896263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.896278 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.998216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.998255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.998267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.998284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:17 crc kubenswrapper[4715]: I1125 12:10:17.998294 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:17Z","lastTransitionTime":"2025-11-25T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.100756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.100814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.100826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.100847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.100857 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:18Z","lastTransitionTime":"2025-11-25T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.203340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.203689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.203800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.204063 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.204216 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:18Z","lastTransitionTime":"2025-11-25T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.308543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.308598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.308610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.308630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.308642 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:18Z","lastTransitionTime":"2025-11-25T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.411488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.411519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.411528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.411543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.411552 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:18Z","lastTransitionTime":"2025-11-25T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.514729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.514780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.514795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.514813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.514824 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:18Z","lastTransitionTime":"2025-11-25T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.618439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.618487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.618498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.618517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.618531 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:18Z","lastTransitionTime":"2025-11-25T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.693166 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.693239 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.693214 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.693179 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:18 crc kubenswrapper[4715]: E1125 12:10:18.693385 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:18 crc kubenswrapper[4715]: E1125 12:10:18.693507 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:18 crc kubenswrapper[4715]: E1125 12:10:18.693675 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:18 crc kubenswrapper[4715]: E1125 12:10:18.693795 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.722269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.722319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.722331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.722352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.722366 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:18Z","lastTransitionTime":"2025-11-25T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.824885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.824941 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.824953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.824969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.824982 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:18Z","lastTransitionTime":"2025-11-25T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.928118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.928212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.928231 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.928255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:18 crc kubenswrapper[4715]: I1125 12:10:18.928270 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:18Z","lastTransitionTime":"2025-11-25T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.031600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.031684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.031704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.031726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.031742 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.134074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.134136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.134149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.134171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.134207 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.237465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.237512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.237522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.237539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.237549 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.339980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.340018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.340026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.340042 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.340051 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.370024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.370074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.370085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.370125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.370137 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: E1125 12:10:19.385995 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:19Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.390341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.390412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.390429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.390453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.390469 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: E1125 12:10:19.410233 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:19Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.413730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.413771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.413784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.413803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.413816 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: E1125 12:10:19.424386 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:19Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.427913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.427944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.427956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.427970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.427979 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: E1125 12:10:19.442364 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:19Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.446644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.446689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.446698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.446720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.446733 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: E1125 12:10:19.460896 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:19Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:19 crc kubenswrapper[4715]: E1125 12:10:19.461053 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.462668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.462712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.462729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.462747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.462758 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.565722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.565762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.565772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.565787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.565799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.668249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.668325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.668389 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.668423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.668445 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.770712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.770756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.770768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.770787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.770797 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.873789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.873846 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.873859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.873879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.873898 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.976579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.976627 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.976639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.976659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:19 crc kubenswrapper[4715]: I1125 12:10:19.976674 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:19Z","lastTransitionTime":"2025-11-25T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.080123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.080201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.080217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.080237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.080249 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:20Z","lastTransitionTime":"2025-11-25T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.186501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.186554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.186566 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.186583 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.186593 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:20Z","lastTransitionTime":"2025-11-25T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.288951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.288993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.289001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.289018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.289027 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:20Z","lastTransitionTime":"2025-11-25T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.392446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.392488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.392499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.392514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.392524 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:20Z","lastTransitionTime":"2025-11-25T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.495981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.496041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.496053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.496070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.496080 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:20Z","lastTransitionTime":"2025-11-25T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.598475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.598559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.598581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.598615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.598637 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:20Z","lastTransitionTime":"2025-11-25T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.693033 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.693064 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:20 crc kubenswrapper[4715]: E1125 12:10:20.693165 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.693215 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.693228 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:20 crc kubenswrapper[4715]: E1125 12:10:20.693287 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:20 crc kubenswrapper[4715]: E1125 12:10:20.693339 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:20 crc kubenswrapper[4715]: E1125 12:10:20.693385 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.700828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.700871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.700897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.700913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.700924 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:20Z","lastTransitionTime":"2025-11-25T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.708749 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.721283 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.734259 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.744412 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.757636 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.769689 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.787245 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.799256 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.803276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.803316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.803329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.803348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.803360 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:20Z","lastTransitionTime":"2025-11-25T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.808821 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.820462 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.831069 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.842956 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.859763 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.871756 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.883773 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.899018 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.905921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.905987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.905998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.906017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.906032 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:20Z","lastTransitionTime":"2025-11-25T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:20 crc kubenswrapper[4715]: I1125 12:10:20.914728 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:20Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.009319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.009385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.009396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.009416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.009428 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.112129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.112202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.112215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.112235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.112247 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.214625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.214671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.214680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.214699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.214712 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.317600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.317662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.317674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.317698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.317712 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.420428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.420495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.420512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.420533 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.420548 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.523712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.523768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.523779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.523801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.523815 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.627238 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.627293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.627305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.627324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.627337 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.730332 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.730391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.730419 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.730443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.730455 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.833569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.833617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.833629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.833648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.833661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.936909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.936989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.937001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.937022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:21 crc kubenswrapper[4715]: I1125 12:10:21.937036 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:21Z","lastTransitionTime":"2025-11-25T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.040275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.040341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.040357 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.040403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.040424 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.142997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.143047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.143058 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.143076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.143096 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.249556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.249622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.249634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.249652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.249667 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.352324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.352383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.352397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.352415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.352428 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.455414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.455475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.455493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.455516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.455531 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.558853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.558907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.558918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.558939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.558952 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.661632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.661691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.661703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.661720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.661748 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.694065 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.694149 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.694171 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.694519 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:22 crc kubenswrapper[4715]: E1125 12:10:22.694611 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:22 crc kubenswrapper[4715]: E1125 12:10:22.694716 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:22 crc kubenswrapper[4715]: E1125 12:10:22.694431 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:22 crc kubenswrapper[4715]: E1125 12:10:22.694821 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.764097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.764151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.764164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.764463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.764481 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.866972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.867052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.867076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.867116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.867139 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.969772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.969818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.969829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.969846 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:22 crc kubenswrapper[4715]: I1125 12:10:22.969856 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:22Z","lastTransitionTime":"2025-11-25T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.072473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.072557 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.072581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.072644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.072661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.173118 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:23 crc kubenswrapper[4715]: E1125 12:10:23.173467 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:10:23 crc kubenswrapper[4715]: E1125 12:10:23.173608 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs podName:3874a0dc-1a53-4587-bb14-7374a9715ae5 nodeName:}" failed. No retries permitted until 2025-11-25 12:10:55.173568395 +0000 UTC m=+105.681071476 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs") pod "network-metrics-daemon-c8s6d" (UID: "3874a0dc-1a53-4587-bb14-7374a9715ae5") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.175450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.175491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.175504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.175524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.175536 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.278209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.278296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.278307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.278324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.278334 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.381094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.381174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.381228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.381254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.381273 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.483743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.483793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.483802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.483820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.483830 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.586922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.586991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.587005 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.587026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.587043 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.690366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.690422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.690504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.690529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.690547 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.793920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.793967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.793981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.794000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.794009 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.896207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.896266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.896277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.896297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.896313 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.999306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.999378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.999397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.999423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:23 crc kubenswrapper[4715]: I1125 12:10:23.999440 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:23Z","lastTransitionTime":"2025-11-25T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.102156 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.102223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.102232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.102251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.102266 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:24Z","lastTransitionTime":"2025-11-25T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.205292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.205339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.205349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.205367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.205377 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:24Z","lastTransitionTime":"2025-11-25T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.307991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.308036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.308048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.308066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.308079 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:24Z","lastTransitionTime":"2025-11-25T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.410560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.410626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.410645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.410672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.410689 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:24Z","lastTransitionTime":"2025-11-25T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.513293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.513373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.513391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.513797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.513857 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:24Z","lastTransitionTime":"2025-11-25T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.617775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.617845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.617861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.617895 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.617910 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:24Z","lastTransitionTime":"2025-11-25T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.693851 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:24 crc kubenswrapper[4715]: E1125 12:10:24.694091 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.694478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.694531 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.694478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:24 crc kubenswrapper[4715]: E1125 12:10:24.694658 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:24 crc kubenswrapper[4715]: E1125 12:10:24.694749 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:24 crc kubenswrapper[4715]: E1125 12:10:24.694868 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.720217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.720266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.720276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.720293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.720304 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:24Z","lastTransitionTime":"2025-11-25T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.823637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.823696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.823714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.823737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.823753 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:24Z","lastTransitionTime":"2025-11-25T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.926642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.926672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.926681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.926697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:24 crc kubenswrapper[4715]: I1125 12:10:24.926705 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:24Z","lastTransitionTime":"2025-11-25T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.030670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.030729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.030745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.030769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.030785 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.134307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.134340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.134349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.134366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.134377 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.237707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.237770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.237789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.237815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.237833 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.340340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.340428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.340441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.340463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.340475 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.443563 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.443642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.443668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.443709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.443734 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.547160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.547243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.547296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.547316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.547331 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.650947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.651001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.651012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.651037 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.651053 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.756116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.756162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.756170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.756206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.756220 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.859981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.860046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.860056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.860073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.860083 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.963382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.963439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.963476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.963509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:25 crc kubenswrapper[4715]: I1125 12:10:25.963535 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:25Z","lastTransitionTime":"2025-11-25T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.065970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.065999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.066007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.066023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.066032 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.168908 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.168953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.168962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.168976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.168987 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.272525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.272587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.272603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.272626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.272641 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.374851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.374887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.374896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.374911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.374921 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.477555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.477635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.477647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.477667 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.477681 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.580815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.580863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.580872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.580891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.580903 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.683727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.683822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.683850 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.683881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.683899 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.693453 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.693515 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.693586 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:26 crc kubenswrapper[4715]: E1125 12:10:26.693781 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.693830 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:26 crc kubenswrapper[4715]: E1125 12:10:26.694109 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:26 crc kubenswrapper[4715]: E1125 12:10:26.694173 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:26 crc kubenswrapper[4715]: E1125 12:10:26.694046 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.786229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.786279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.786288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.786308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.786318 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.891780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.891821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.891830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.891843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.891855 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.994558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.994612 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.994630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.994650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:26 crc kubenswrapper[4715]: I1125 12:10:26.994661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:26Z","lastTransitionTime":"2025-11-25T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.097788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.097835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.097845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.097861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.097872 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:27Z","lastTransitionTime":"2025-11-25T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.200024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.200083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.200096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.200117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.200129 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:27Z","lastTransitionTime":"2025-11-25T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.302765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.302825 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.302842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.302871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.302889 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:27Z","lastTransitionTime":"2025-11-25T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.405909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.405970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.405989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.406053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.406078 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:27Z","lastTransitionTime":"2025-11-25T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.508925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.508999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.509010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.509031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.509047 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:27Z","lastTransitionTime":"2025-11-25T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.612732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.612798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.612812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.612828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.612840 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:27Z","lastTransitionTime":"2025-11-25T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.716081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.716153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.716169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.716224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.716244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:27Z","lastTransitionTime":"2025-11-25T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.818592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.818640 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.818652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.818670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.818681 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:27Z","lastTransitionTime":"2025-11-25T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.922234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.922329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.922360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.922391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:27 crc kubenswrapper[4715]: I1125 12:10:27.922415 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:27Z","lastTransitionTime":"2025-11-25T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.025450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.025504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.025517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.025539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.025552 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.128304 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.128363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.128372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.128388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.128401 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.213606 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/0.log" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.213662 4715 generic.go:334] "Generic (PLEG): container finished" podID="6371ac18-2361-43bb-b474-32bf53713cf5" containerID="eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798" exitCode=1 Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.213695 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5mzr5" event={"ID":"6371ac18-2361-43bb-b474-32bf53713cf5","Type":"ContainerDied","Data":"eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.214109 4715 scope.go:117] "RemoveContainer" containerID="eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.226945 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.232845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.232935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.232948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.232967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.232978 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.243652 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.257646 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.275599 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.288668 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.302919 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.315110 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.327327 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.335731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.335779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.335789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.335806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.335817 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.343922 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.354429 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.366633 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:27Z\\\",\\\"message\\\":\\\"2025-11-25T12:09:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd\\\\n2025-11-25T12:09:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd to /host/opt/cni/bin/\\\\n2025-11-25T12:09:42Z [verbose] multus-daemon started\\\\n2025-11-25T12:09:42Z [verbose] Readiness Indicator file check\\\\n2025-11-25T12:10:27Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.381883 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.390740 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.403912 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.417324 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.431047 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.438216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.438252 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.438262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.438277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.438289 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.444236 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:28Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.540284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.540320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.540330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.540346 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.540357 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.645391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.645454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.645466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.645487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.645500 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.692997 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.693079 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.693164 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:28 crc kubenswrapper[4715]: E1125 12:10:28.693219 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:28 crc kubenswrapper[4715]: E1125 12:10:28.693350 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.693378 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:28 crc kubenswrapper[4715]: E1125 12:10:28.693578 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:28 crc kubenswrapper[4715]: E1125 12:10:28.693666 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.748225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.748303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.748315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.748336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.748350 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.852040 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.852091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.852137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.852161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.852174 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.956169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.956261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.956285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.956311 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:28 crc kubenswrapper[4715]: I1125 12:10:28.956326 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:28Z","lastTransitionTime":"2025-11-25T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.059127 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.059235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.059261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.059292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.059309 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.162348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.162397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.162412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.162427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.162438 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.217908 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/0.log" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.217971 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5mzr5" event={"ID":"6371ac18-2361-43bb-b474-32bf53713cf5","Type":"ContainerStarted","Data":"2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.232031 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.251031 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.266529 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.266950 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.267012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.267025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.267046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.267060 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.282173 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.303220 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.317398 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.331471 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.346818 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.359353 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.370147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.370181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.370211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.370230 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.370244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.372741 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.385643 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.405269 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:27Z\\\",\\\"message\\\":\\\"2025-11-25T12:09:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd\\\\n2025-11-25T12:09:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd to /host/opt/cni/bin/\\\\n2025-11-25T12:09:42Z [verbose] multus-daemon started\\\\n2025-11-25T12:09:42Z [verbose] Readiness Indicator file check\\\\n2025-11-25T12:10:27Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:10:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.427839 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.441979 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.458753 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.472602 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.473210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.473255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.473271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.473294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.473312 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.489466 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.576385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.576442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.576454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.576471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.576480 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.678827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.678914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.678925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.678942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.678954 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.781549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.781653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.781711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.781736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.781752 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.804755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.804825 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.804843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.804873 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.804891 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: E1125 12:10:29.817792 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.821740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.821799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.821814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.821833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.821846 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: E1125 12:10:29.842265 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.846902 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.846967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.846980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.847001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.847015 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: E1125 12:10:29.862360 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.867504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.867541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.867551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.867564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.867574 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: E1125 12:10:29.882817 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.886978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.887020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.887030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.887046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.887058 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:29 crc kubenswrapper[4715]: E1125 12:10:29.902836 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:29Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:29 crc kubenswrapper[4715]: E1125 12:10:29.903002 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.905478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.905515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.905523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.905561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:29 crc kubenswrapper[4715]: I1125 12:10:29.905574 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:29Z","lastTransitionTime":"2025-11-25T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.008282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.008334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.008350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.008370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.008386 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.111281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.111364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.111374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.111390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.111400 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.214908 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.214953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.214966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.214982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.214994 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.317364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.317404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.317413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.317428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.317437 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.420635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.420691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.420704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.420723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.420735 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.525116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.525168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.525195 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.525217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.525229 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.628548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.628598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.628633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.628649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.628658 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.692916 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.693008 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.693040 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.693118 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:30 crc kubenswrapper[4715]: E1125 12:10:30.693116 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:30 crc kubenswrapper[4715]: E1125 12:10:30.693954 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:30 crc kubenswrapper[4715]: E1125 12:10:30.694047 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:30 crc kubenswrapper[4715]: E1125 12:10:30.694085 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.723327 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.734992 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.735026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.735035 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.735050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.735061 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.746739 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.765364 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.786851 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.804600 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.827469 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.839278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.839351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.839373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.839403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.839423 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.846633 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.862722 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.877540 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.893645 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.911871 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.926396 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.948085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.948163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.948176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.948233 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.948250 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:30Z","lastTransitionTime":"2025-11-25T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.950418 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.965437 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:30 crc kubenswrapper[4715]: I1125 12:10:30.990934 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:27Z\\\",\\\"message\\\":\\\"2025-11-25T12:09:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd\\\\n2025-11-25T12:09:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd to /host/opt/cni/bin/\\\\n2025-11-25T12:09:42Z [verbose] multus-daemon started\\\\n2025-11-25T12:09:42Z [verbose] Readiness Indicator file check\\\\n2025-11-25T12:10:27Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:10:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:30Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.009986 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:31Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.023758 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:31Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.051391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.051465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.051481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.051503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.051521 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.154782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.154874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.154899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.154944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.154979 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.258709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.258768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.258783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.258801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.258812 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.361636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.361674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.361685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.361703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.361714 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.464422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.464471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.464482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.464498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.464511 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.567175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.567229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.567240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.567256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.567266 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.670126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.670160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.670169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.670196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.670206 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.772959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.773014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.773027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.773047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.773062 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.876553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.876650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.876669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.876694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.876711 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.979164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.979245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.979261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.979284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:31 crc kubenswrapper[4715]: I1125 12:10:31.979301 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:31Z","lastTransitionTime":"2025-11-25T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.082639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.082688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.082700 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.082717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.082730 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:32Z","lastTransitionTime":"2025-11-25T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.185429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.185492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.185505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.185524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.185537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:32Z","lastTransitionTime":"2025-11-25T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.288623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.288689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.288699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.288715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.288728 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:32Z","lastTransitionTime":"2025-11-25T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.394774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.394803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.394812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.394825 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.394834 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:32Z","lastTransitionTime":"2025-11-25T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.497521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.497568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.497581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.497599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.497615 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:32Z","lastTransitionTime":"2025-11-25T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.604448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.604491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.604502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.604523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.604535 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:32Z","lastTransitionTime":"2025-11-25T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.693384 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.693458 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.693478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.693534 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:32 crc kubenswrapper[4715]: E1125 12:10:32.693561 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:32 crc kubenswrapper[4715]: E1125 12:10:32.693752 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:32 crc kubenswrapper[4715]: E1125 12:10:32.693809 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:32 crc kubenswrapper[4715]: E1125 12:10:32.693880 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.694525 4715 scope.go:117] "RemoveContainer" containerID="d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.708173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.708404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.708522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.708633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.708729 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:32Z","lastTransitionTime":"2025-11-25T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.811597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.812048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.812078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.812109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.812132 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:32Z","lastTransitionTime":"2025-11-25T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.915449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.915490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.915504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.915523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:32 crc kubenswrapper[4715]: I1125 12:10:32.915537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:32Z","lastTransitionTime":"2025-11-25T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.018917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.018958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.018979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.018999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.019013 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.122146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.122209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.122228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.122249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.122262 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.224883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.224933 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.224949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.224966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.224977 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.234644 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/2.log" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.237656 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.238086 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.257107 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.270947 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.285425 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.298472 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.309144 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.322505 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:27Z\\\",\\\"message\\\":\\\"2025-11-25T12:09:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd\\\\n2025-11-25T12:09:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd to /host/opt/cni/bin/\\\\n2025-11-25T12:09:42Z [verbose] multus-daemon started\\\\n2025-11-25T12:09:42Z [verbose] Readiness Indicator file check\\\\n2025-11-25T12:10:27Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:10:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.326670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.326713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.326724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.326741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.326753 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.341083 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.352206 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.368362 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.381491 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.394455 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.410130 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.425726 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.431146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.431227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.431246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.431273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.431292 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.440686 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.455867 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.473997 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.490419 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.534576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.534631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.534648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.534671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.534688 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.637964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.638024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.638042 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.638067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.638085 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.740394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.740431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.740442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.740458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.740470 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.843395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.843446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.843458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.843475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.843486 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.946499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.946561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.946575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.946595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:33 crc kubenswrapper[4715]: I1125 12:10:33.946611 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:33Z","lastTransitionTime":"2025-11-25T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.049809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.049873 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.049885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.049905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.049917 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.152747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.152813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.152830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.152854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.152874 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.245277 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/3.log" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.246170 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/2.log" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.250281 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" exitCode=1 Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.250351 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.250410 4715 scope.go:117] "RemoveContainer" containerID="d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.251631 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:10:34 crc kubenswrapper[4715]: E1125 12:10:34.251893 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.268537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.268631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.268719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.268817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.268844 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.286548 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:27Z\\\",\\\"message\\\":\\\"2025-11-25T12:09:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd\\\\n2025-11-25T12:09:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd to /host/opt/cni/bin/\\\\n2025-11-25T12:09:42Z [verbose] multus-daemon started\\\\n2025-11-25T12:09:42Z [verbose] Readiness Indicator file check\\\\n2025-11-25T12:10:27Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:10:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.315709 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d99fb0ee227ec9ef5fd54077cca55ce166ca428dcff362c9c1e22b809af8a386\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:04Z\\\",\\\"message\\\":\\\"machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 12:10:04.556382 6376 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:33Z\\\",\\\"message\\\":\\\" Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z]\\\\nI1125 12:10:33.509922 6766 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tzl95\\\\nI1125 12:10:33.509921 6766 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-5mzr5\\\\nI1125 12:10:33.509930 6766 services_controller.go:360] Finished syncing service router-internal-default on namespace openshift-ingress for network=default : 3.219058ms\\\\nI1125 12:10:33.509934 6766 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tzl95 in node crc\\\\nI1125 12:10:33.509907 6766 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-qw7ss\\\\nI1125 12:10:33.509899 6766 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Bala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.328325 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.345821 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.361083 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.371079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.371124 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.371137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.371153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.371165 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.372980 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.382261 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.393071 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.407476 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.418903 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.434135 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.447472 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.459975 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.471954 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.473881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.473946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.473969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.473998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.474019 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.484868 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.496307 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.508885 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:34Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.577403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.577436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.577446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.577462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.577474 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.680441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.680513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.680547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.680567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.680580 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.693106 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.693161 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.693257 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:34 crc kubenswrapper[4715]: E1125 12:10:34.693301 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.693429 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:34 crc kubenswrapper[4715]: E1125 12:10:34.693510 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:34 crc kubenswrapper[4715]: E1125 12:10:34.693534 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:34 crc kubenswrapper[4715]: E1125 12:10:34.693580 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.783052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.783111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.783132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.783168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.783222 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.887137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.887210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.887223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.887242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.887256 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.990358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.990404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.990414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.990432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:34 crc kubenswrapper[4715]: I1125 12:10:34.990445 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:34Z","lastTransitionTime":"2025-11-25T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.093023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.093067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.093082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.093099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.093112 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:35Z","lastTransitionTime":"2025-11-25T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.195880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.195925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.195938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.195956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.195968 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:35Z","lastTransitionTime":"2025-11-25T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.262117 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/3.log" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.267440 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:10:35 crc kubenswrapper[4715]: E1125 12:10:35.267744 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.290494 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.298624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.298693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.298708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.298734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.298753 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:35Z","lastTransitionTime":"2025-11-25T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.306346 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.323119 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.339176 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.355706 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.370676 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.384752 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.401764 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.402681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.402706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.402715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.402728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.402740 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:35Z","lastTransitionTime":"2025-11-25T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.413964 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.429049 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.443776 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.457341 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.473282 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.488501 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.500799 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.504678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.504716 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.504734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.504751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.504762 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:35Z","lastTransitionTime":"2025-11-25T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.517922 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:27Z\\\",\\\"message\\\":\\\"2025-11-25T12:09:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd\\\\n2025-11-25T12:09:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd to /host/opt/cni/bin/\\\\n2025-11-25T12:09:42Z [verbose] multus-daemon started\\\\n2025-11-25T12:09:42Z [verbose] Readiness Indicator file check\\\\n2025-11-25T12:10:27Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:10:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.537612 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:33Z\\\",\\\"message\\\":\\\" Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z]\\\\nI1125 12:10:33.509922 6766 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tzl95\\\\nI1125 12:10:33.509921 6766 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-5mzr5\\\\nI1125 12:10:33.509930 6766 services_controller.go:360] Finished syncing service router-internal-default on namespace openshift-ingress for network=default : 3.219058ms\\\\nI1125 12:10:33.509934 6766 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tzl95 in node crc\\\\nI1125 12:10:33.509907 6766 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-qw7ss\\\\nI1125 12:10:33.509899 6766 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Bala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:35Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.607076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.607144 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.607161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.607217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.607236 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:35Z","lastTransitionTime":"2025-11-25T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.710754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.710802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.710812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.710828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.710838 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:35Z","lastTransitionTime":"2025-11-25T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.813454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.813506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.813518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.813535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.813550 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:35Z","lastTransitionTime":"2025-11-25T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.916729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.916817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.916830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.916849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:35 crc kubenswrapper[4715]: I1125 12:10:35.916862 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:35Z","lastTransitionTime":"2025-11-25T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.019025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.019079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.019102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.019133 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.019155 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.121753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.121805 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.121815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.121828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.121838 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.224416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.224463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.224476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.224492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.224504 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.327399 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.327433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.327444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.327460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.327472 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.430759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.430821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.430838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.430863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.430882 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.512689 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.512881 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.512981 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:11:40.51296094 +0000 UTC m=+151.020463961 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.534206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.534270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.534282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.534300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.534316 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.613282 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.613423 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613467 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:40.613440155 +0000 UTC m=+151.120943176 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.613552 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.613587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613594 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613623 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613642 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613703 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613714 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 12:11:40.613691602 +0000 UTC m=+151.121194663 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613718 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613754 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613755 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613785 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 12:11:40.613778744 +0000 UTC m=+151.121281765 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.613828 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 12:11:40.613810645 +0000 UTC m=+151.121313746 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.636852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.636901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.636920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.636942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.636958 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.693203 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.693250 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.693340 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.693486 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.693528 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.693657 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.693732 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:36 crc kubenswrapper[4715]: E1125 12:10:36.693810 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.738901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.738952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.738961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.738977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.738988 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.841329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.841415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.841433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.841450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.841462 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.944107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.944156 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.944168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.944422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:36 crc kubenswrapper[4715]: I1125 12:10:36.944439 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:36Z","lastTransitionTime":"2025-11-25T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.053537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.053585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.053595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.053609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.053619 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.156422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.156470 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.156480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.156497 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.156509 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.260667 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.260767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.260794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.260862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.260885 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.363660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.363696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.363706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.363720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.363729 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.466463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.466505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.466517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.466530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.466539 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.568973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.569038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.569051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.569067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.569126 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.672611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.672653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.672675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.672689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.672698 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.775359 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.775428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.775446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.775467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.775511 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.879065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.879141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.879166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.879247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.879274 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.982921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.982987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.983002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.983022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:37 crc kubenswrapper[4715]: I1125 12:10:37.983038 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:37Z","lastTransitionTime":"2025-11-25T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.085853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.085896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.085906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.085924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.085936 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:38Z","lastTransitionTime":"2025-11-25T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.188977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.189022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.189036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.189056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.189068 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:38Z","lastTransitionTime":"2025-11-25T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.293041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.293090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.293102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.293123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.293136 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:38Z","lastTransitionTime":"2025-11-25T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.395691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.395746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.395759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.395777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.395790 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:38Z","lastTransitionTime":"2025-11-25T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.498614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.498664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.498678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.498694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.498704 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:38Z","lastTransitionTime":"2025-11-25T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.601779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.601813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.601824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.601837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.601847 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:38Z","lastTransitionTime":"2025-11-25T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.693356 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.693489 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:38 crc kubenswrapper[4715]: E1125 12:10:38.693493 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.693529 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:38 crc kubenswrapper[4715]: E1125 12:10:38.693562 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.693368 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:38 crc kubenswrapper[4715]: E1125 12:10:38.693623 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:38 crc kubenswrapper[4715]: E1125 12:10:38.693669 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.704699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.704752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.704765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.704786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.704799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:38Z","lastTransitionTime":"2025-11-25T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.807429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.807480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.807493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.807513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.807525 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:38Z","lastTransitionTime":"2025-11-25T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.910384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.910428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.910440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.910457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:38 crc kubenswrapper[4715]: I1125 12:10:38.910472 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:38Z","lastTransitionTime":"2025-11-25T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.012793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.012847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.012859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.012877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.012888 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.115531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.115600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.115608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.115621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.115630 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.218170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.218237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.218250 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.218266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.218279 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.321848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.321905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.321919 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.321943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.321960 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.424458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.424515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.424529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.424548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.424564 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.526713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.526773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.526784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.526803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.526815 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.629311 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.629354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.629368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.629383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.629394 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.710295 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.732165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.732221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.732229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.732243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.732251 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.835116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.835477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.835558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.835627 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.835714 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.908469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.908541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.908554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.908574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.908589 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: E1125 12:10:39.925661 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.930487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.930572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.930600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.930637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.930663 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: E1125 12:10:39.946515 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.950268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.950310 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.950320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.950336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.950348 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: E1125 12:10:39.965301 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.969695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.969742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.969755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.969774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.969786 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:39 crc kubenswrapper[4715]: E1125 12:10:39.986284 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:39Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.990956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.991031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.991054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.991086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:39 crc kubenswrapper[4715]: I1125 12:10:39.991110 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:39Z","lastTransitionTime":"2025-11-25T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: E1125 12:10:40.009989 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b3d33238-179c-4051-9790-e67bba33905c\\\",\\\"systemUUID\\\":\\\"aa0ef64c-25d2-4dc4-b8f3-98dcad372584\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: E1125 12:10:40.010112 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.012261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.012303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.012316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.012333 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.012343 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.115551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.115616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.115636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.115661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.115678 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.218955 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.219012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.219024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.219041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.219056 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.322000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.322043 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.322055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.322070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.322081 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.424860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.425580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.425626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.425646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.425658 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.528799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.528883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.528898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.528917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.528931 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.631790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.631868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.631881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.631903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.631918 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.693583 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.693826 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.693861 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.693843 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:40 crc kubenswrapper[4715]: E1125 12:10:40.693973 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:40 crc kubenswrapper[4715]: E1125 12:10:40.694093 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:40 crc kubenswrapper[4715]: E1125 12:10:40.694218 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:40 crc kubenswrapper[4715]: E1125 12:10:40.694346 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.709516 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.710522 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.727085 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.734648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.734708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.734727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.734751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.734805 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.740245 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9638dd-8f44-43b9-b801-4ffbc3941b79\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fb89eaac0a61c330c8bb3fcb44112ae9b50bb54ae37e2b3a4eec3b72b9e3af1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf2c3e3c314d9bd3a6ff94698869fb22378770f84ae80b6c2c80ce410d9b7c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5da263adc4d612f790ea651741219b7e06c2943ba885a473acfe14787f5dc451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab6a39f01b5e160134f4015a689cdfe93471794555cd223b8bcfedf235ee783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.754559 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ed7fbee88aa72e79d8be07837c128cca702b224a2ac610273781287618fec43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.768531 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.785132 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26eb70b2cee9e3c9a3a97f49e9c08d6f25b70c8e5a425e9a1361a566864b1ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.801378 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51676837-adc4-4424-b527-920a6528b6a2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ca4935f3ed94dc310e625300644e7924a09bbffe6bd447a0e6f499b412cf99d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dk9f9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.820367 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8ba0cf3-8a5a-4282-a6f0-c5119a7cff2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eec07c89c23392b169631f5f9158638860b55f3d28f8020544de9f37c60a4f3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ea15fda593ef5fa0e955208653246431795d869a173586f5f351487210654af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbb2632d8f847b942f17892b26bef65ee3e48f5aed708a78f3d0d8983fb71fb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e86a083928b448c9259e8509e05d79e507060f42f2b3eb921736b9843fbfeaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fa34a62aecc9004e8d7b2b7ba617af825eb65f25a41985731cfc188371e6ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a3a6f05a92f7f26d9544337b2f3e60473ed75df318cd06050f927d67795b399\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://469c8a7c85999edf4c1de31eb3d1f74bec8837de979439e5315a0a4d46aa3df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9922\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5fwzm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.833512 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rncj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92f2a401-5b59-48a9-8e15-35d5f62afd66\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd8cbf4226ce47f80f6b67ee8d989f8187944b5a758755867248d9716f2ea452\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x8ndn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rncj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.837438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.837495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.837511 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.837535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.837552 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.846280 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ea83ae-7f5a-4224-8e39-c2a86950ee7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1664044f0e5c7794dd6cb5b35a0f2231f69d4bcc8dfb359432aa724420f6172b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8079ecca7e922b986b5f5185ea09245df6d41f773c1e88b13f1de0e611b6534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6t9m5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vwc5f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.866977 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c22559b8-7b65-4364-ab7a-b6ddfe67e7a4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df6d9273ca8652c52189a435aa4b06d724fa18dbca2b78141e63e695b7f6f3d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fde5268d02a6206f73d1a64b77d7145092e59758bc01c80e096b9e12d3f78d3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b42e0d2a865c7b7b7ae6efedbcbe6f9ca83e8f71c2334caf3c24a587acd1a574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4fe863afe95d6e4d7bfbd829de3bd73e2fc041f2cc75aae83922cd063a102fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ade9a469a8f08e584b0d95b466bd612154a082b1487eebb198739c1b198c595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39c7a80a7f8bf43652f7ed2c89123140f7382ec852049e9425458acec423a0b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39c7a80a7f8bf43652f7ed2c89123140f7382ec852049e9425458acec423a0b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f417bb62facafa423eac71cfdc98e439ca08948cae893d49dda30dddbe57e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39f417bb62facafa423eac71cfdc98e439ca08948cae893d49dda30dddbe57e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://258869dd8353aa6544618e6c080fba6d21023f16576ea92dd55d74faf180fa7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://258869dd8353aa6544618e6c080fba6d21023f16576ea92dd55d74faf180fa7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.881100 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2e09878-9356-4e26-8774-73369b455905\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 12:09:32.091308 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 12:09:32.091498 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 12:09:32.092393 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1264125343/tls.crt::/tmp/serving-cert-1264125343/tls.key\\\\\\\"\\\\nI1125 12:09:32.500396 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 12:09:32.508078 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 12:09:32.508119 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 12:09:32.508158 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 12:09:32.508170 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 12:09:32.518018 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1125 12:09:32.518039 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1125 12:09:32.518062 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518070 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 12:09:32.518077 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 12:09:32.518082 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 12:09:32.518086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 12:09:32.518090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1125 12:09:32.521256 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:16Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.893642 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f0bf5a0-301e-4259-b197-28ae3360374c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://145794e939c1c0185ee88e6406234e880007421a8d2c5a20ea2066fdf2c53a25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31803bc0e120ca90a0617b4594647be14faa970b9c916794b2d9df3452b04c1f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a88275d89e775ec428c6ec3ac66972ab55af06da0cbe07b29f949ea20ef38040\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.905957 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3e62e221a189b504c0a3df7b7a65ea6816328ccaf724e45202c2b5b11393c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dc6ed384d17a07afad48571faebc37f50b6949952673f5fb247599328d6c4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.915735 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qw7ss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6ba7789-d2b7-46eb-97fd-0f61f98e8e15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db95e2feca51c9263e53151318f5c5419923c045f63bb3adada15f1d910f72d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxcrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qw7ss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.927735 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5mzr5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6371ac18-2361-43bb-b474-32bf53713cf5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:27Z\\\",\\\"message\\\":\\\"2025-11-25T12:09:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd\\\\n2025-11-25T12:09:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6ad17a6a-7507-450f-9819-9bd1d17a28cd to /host/opt/cni/bin/\\\\n2025-11-25T12:09:42Z [verbose] multus-daemon started\\\\n2025-11-25T12:09:42Z [verbose] Readiness Indicator file check\\\\n2025-11-25T12:10:27Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:10:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rw5xd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5mzr5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.940073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.940111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.940126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.940144 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.940158 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:40Z","lastTransitionTime":"2025-11-25T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.947453 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02385935-e23a-4708-b57f-fd393d05651b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T12:10:33Z\\\",\\\"message\\\":\\\" Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:33Z is after 2025-08-24T17:21:41Z]\\\\nI1125 12:10:33.509922 6766 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tzl95\\\\nI1125 12:10:33.509921 6766 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-5mzr5\\\\nI1125 12:10:33.509930 6766 services_controller.go:360] Finished syncing service router-internal-default on namespace openshift-ingress for network=default : 3.219058ms\\\\nI1125 12:10:33.509934 6766 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tzl95 in node crc\\\\nI1125 12:10:33.509907 6766 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-qw7ss\\\\nI1125 12:10:33.509899 6766 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Bala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T12:10:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T12:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T12:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T12:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kpqs9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tzl95\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:40 crc kubenswrapper[4715]: I1125 12:10:40.960768 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3874a0dc-1a53-4587-bb14-7374a9715ae5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T12:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cflzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T12:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-c8s6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T12:10:40Z is after 2025-08-24T17:21:41Z" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.043489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.043534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.043543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.043560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.043570 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.146922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.146979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.146989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.147007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.147020 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.249474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.249509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.249520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.249534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.249545 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.352593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.352637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.352646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.352661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.352671 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.455286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.455356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.455381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.455409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.455429 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.557700 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.557746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.557757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.557774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.557786 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.660031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.660094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.660111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.660129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.660170 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.762911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.763006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.763032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.763065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.763080 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.865900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.865954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.865966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.865983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.865995 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.969016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.969062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.969071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.969090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:41 crc kubenswrapper[4715]: I1125 12:10:41.969102 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:41Z","lastTransitionTime":"2025-11-25T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.072493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.072551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.072561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.072578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.072590 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:42Z","lastTransitionTime":"2025-11-25T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.175523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.175572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.175590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.175607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.175618 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:42Z","lastTransitionTime":"2025-11-25T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.277745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.277808 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.277827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.277844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.277855 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:42Z","lastTransitionTime":"2025-11-25T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.380322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.380383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.380396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.380415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.380429 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:42Z","lastTransitionTime":"2025-11-25T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.484061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.484669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.484679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.484694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.484703 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:42Z","lastTransitionTime":"2025-11-25T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.587764 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.587804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.587813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.587828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.587839 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:42Z","lastTransitionTime":"2025-11-25T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.691702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.691779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.691801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.691828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.691852 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:42Z","lastTransitionTime":"2025-11-25T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.692949 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.692949 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:42 crc kubenswrapper[4715]: E1125 12:10:42.693060 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.693102 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:42 crc kubenswrapper[4715]: E1125 12:10:42.693222 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:42 crc kubenswrapper[4715]: E1125 12:10:42.693268 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.693381 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:42 crc kubenswrapper[4715]: E1125 12:10:42.693450 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.794912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.794960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.794968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.794984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.794994 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:42Z","lastTransitionTime":"2025-11-25T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.898029 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.898094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.898105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.898122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:42 crc kubenswrapper[4715]: I1125 12:10:42.898138 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:42Z","lastTransitionTime":"2025-11-25T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.001719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.001781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.001799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.001825 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.001848 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.103968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.104018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.104027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.104042 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.104054 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.207358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.207409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.207424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.207443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.207457 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.309737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.309784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.309800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.309821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.309838 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.412549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.412628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.412646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.412668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.412685 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.515753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.515842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.515861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.515890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.515908 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.620067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.620128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.620143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.620166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.620179 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.722576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.722612 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.722625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.722649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.722661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.825673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.825737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.825749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.825770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.825782 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.928703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.928789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.928804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.928829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:43 crc kubenswrapper[4715]: I1125 12:10:43.928844 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:43Z","lastTransitionTime":"2025-11-25T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.031722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.031765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.031778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.031797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.031810 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.134974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.135047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.135062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.135093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.135108 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.238360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.238400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.238411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.238428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.238440 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.341300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.341657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.341673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.341690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.341701 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.443586 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.443620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.443631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.443647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.443658 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.546645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.546717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.546740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.546767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.546785 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.649706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.649754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.649770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.649786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.649797 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.693795 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.693868 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.693859 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:44 crc kubenswrapper[4715]: E1125 12:10:44.693993 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.693822 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:44 crc kubenswrapper[4715]: E1125 12:10:44.694110 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:44 crc kubenswrapper[4715]: E1125 12:10:44.694360 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:44 crc kubenswrapper[4715]: E1125 12:10:44.694478 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.752040 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.752105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.752115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.752137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.752150 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.855800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.855866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.855877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.855897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.855912 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.960783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.960838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.960852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.960867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:44 crc kubenswrapper[4715]: I1125 12:10:44.960878 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:44Z","lastTransitionTime":"2025-11-25T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.064698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.064753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.064763 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.064783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.064794 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.171974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.172062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.172076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.172098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.172117 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.275102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.275166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.275213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.275236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.275251 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.377323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.377372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.377384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.377401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.377411 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.481051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.481104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.481113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.481130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.481140 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.584140 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.584215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.584226 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.584242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.584253 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.687458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.687500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.687508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.687521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.687530 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.789968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.790009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.790019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.790033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.790045 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.892049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.892105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.892120 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.892136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.892147 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.995483 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.995549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.995568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.995592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:45 crc kubenswrapper[4715]: I1125 12:10:45.995610 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:45Z","lastTransitionTime":"2025-11-25T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.097761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.097802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.097813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.097827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.097839 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:46Z","lastTransitionTime":"2025-11-25T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.201362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.201431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.201441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.201464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.201482 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:46Z","lastTransitionTime":"2025-11-25T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.304696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.304757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.304776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.304803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.304817 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:46Z","lastTransitionTime":"2025-11-25T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.407513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.407554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.407565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.407578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.407587 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:46Z","lastTransitionTime":"2025-11-25T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.510461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.510671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.510796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.510917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.510982 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:46Z","lastTransitionTime":"2025-11-25T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.614635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.614691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.614705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.614724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.614738 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:46Z","lastTransitionTime":"2025-11-25T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.693165 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.693226 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.693422 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:46 crc kubenswrapper[4715]: E1125 12:10:46.693608 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.693892 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:46 crc kubenswrapper[4715]: E1125 12:10:46.694010 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:46 crc kubenswrapper[4715]: E1125 12:10:46.694233 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:46 crc kubenswrapper[4715]: E1125 12:10:46.694421 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.717076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.717145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.717172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.717258 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.717290 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:46Z","lastTransitionTime":"2025-11-25T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.820430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.820492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.820514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.820548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.820572 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:46Z","lastTransitionTime":"2025-11-25T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.923340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.923391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.923405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.923424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:46 crc kubenswrapper[4715]: I1125 12:10:46.923436 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:46Z","lastTransitionTime":"2025-11-25T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.026398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.026434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.026443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.026457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.026466 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.130057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.130109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.130126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.130150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.130167 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.233118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.233249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.233268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.233328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.233347 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.335503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.335562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.335579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.335601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.335620 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.437921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.437968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.437977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.437993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.438003 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.541975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.542038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.542048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.542084 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.542096 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.645088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.645160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.645210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.645242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.645260 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.747441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.747483 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.747497 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.747511 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.747521 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.849914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.849956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.849966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.849987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.849998 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.953495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.953542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.953555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.953571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:47 crc kubenswrapper[4715]: I1125 12:10:47.953584 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:47Z","lastTransitionTime":"2025-11-25T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.056340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.056656 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.056727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.056819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.056882 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.159713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.159759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.159771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.159786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.159799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.262575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.262616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.262625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.262639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.262648 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.365326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.365370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.365382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.365404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.365416 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.468562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.468649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.468667 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.468687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.468734 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.570944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.571293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.571376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.571458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.571551 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.674436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.674486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.674498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.674513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.674521 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.693884 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.693937 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.693904 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.694022 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:48 crc kubenswrapper[4715]: E1125 12:10:48.694111 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:48 crc kubenswrapper[4715]: E1125 12:10:48.694228 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:48 crc kubenswrapper[4715]: E1125 12:10:48.694285 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:48 crc kubenswrapper[4715]: E1125 12:10:48.694483 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.776938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.777605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.777649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.777668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.777681 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.881093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.881149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.881160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.881177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.881207 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.983897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.983959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.983975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.983995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:48 crc kubenswrapper[4715]: I1125 12:10:48.984008 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:48Z","lastTransitionTime":"2025-11-25T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.086686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.086734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.086745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.086759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.086768 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:49Z","lastTransitionTime":"2025-11-25T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.189116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.189152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.189180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.189224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.189270 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:49Z","lastTransitionTime":"2025-11-25T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.292084 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.292145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.292154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.292171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.292211 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:49Z","lastTransitionTime":"2025-11-25T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.394363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.394416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.394429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.394446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.394458 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:49Z","lastTransitionTime":"2025-11-25T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.497458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.497493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.497502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.497515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.497523 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:49Z","lastTransitionTime":"2025-11-25T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.600790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.600874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.600892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.600911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.600924 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:49Z","lastTransitionTime":"2025-11-25T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.703158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.703213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.703227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.703244 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.703255 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:49Z","lastTransitionTime":"2025-11-25T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.805849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.805886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.805897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.805913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.805924 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:49Z","lastTransitionTime":"2025-11-25T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.908489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.908522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.908573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.908589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:49 crc kubenswrapper[4715]: I1125 12:10:49.908598 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:49Z","lastTransitionTime":"2025-11-25T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.012836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.012892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.012909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.012929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.012943 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:50Z","lastTransitionTime":"2025-11-25T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.115999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.116061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.116079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.116103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.116123 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:50Z","lastTransitionTime":"2025-11-25T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.219286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.219331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.219344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.219361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.219373 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:50Z","lastTransitionTime":"2025-11-25T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.322036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.322111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.322134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.322161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.322217 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:50Z","lastTransitionTime":"2025-11-25T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.382362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.382418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.382429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.382443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.382451 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T12:10:50Z","lastTransitionTime":"2025-11-25T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.425338 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p"] Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.425716 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.427919 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.427980 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.433224 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.433396 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.454804 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vwc5f" podStartSLOduration=73.45478161 podStartE2EDuration="1m13.45478161s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.454016789 +0000 UTC m=+100.961519810" watchObservedRunningTime="2025-11-25 12:10:50.45478161 +0000 UTC m=+100.962284631" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.455017 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-rncj5" podStartSLOduration=74.455009616 podStartE2EDuration="1m14.455009616s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.442877815 +0000 UTC m=+100.950380836" watchObservedRunningTime="2025-11-25 12:10:50.455009616 +0000 UTC m=+100.962512637" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.463642 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.463787 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.463883 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.463952 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.463994 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.492731 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-5mzr5" podStartSLOduration=74.492712616 podStartE2EDuration="1m14.492712616s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.468534926 +0000 UTC m=+100.976037947" watchObservedRunningTime="2025-11-25 12:10:50.492712616 +0000 UTC m=+101.000215637" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.526309 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=11.526290423 podStartE2EDuration="11.526290423s" podCreationTimestamp="2025-11-25 12:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.525402849 +0000 UTC m=+101.032905870" watchObservedRunningTime="2025-11-25 12:10:50.526290423 +0000 UTC m=+101.033793444" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.552959 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.552937491 podStartE2EDuration="1m18.552937491s" podCreationTimestamp="2025-11-25 12:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.539477983 +0000 UTC m=+101.046981014" watchObservedRunningTime="2025-11-25 12:10:50.552937491 +0000 UTC m=+101.060440512" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.553111 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.553105986 podStartE2EDuration="1m18.553105986s" podCreationTimestamp="2025-11-25 12:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.552871999 +0000 UTC m=+101.060375030" watchObservedRunningTime="2025-11-25 12:10:50.553105986 +0000 UTC m=+101.060609017" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.565257 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.565328 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.565379 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.565412 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.565472 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.565547 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.565565 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.566577 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.574214 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.580469 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f145f4d6-2e9e-4d9e-afb0-fe334f747b27-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8tv2p\" (UID: \"f145f4d6-2e9e-4d9e-afb0-fe334f747b27\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.596121 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-qw7ss" podStartSLOduration=74.59610005 podStartE2EDuration="1m14.59610005s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.58071381 +0000 UTC m=+101.088216831" watchObservedRunningTime="2025-11-25 12:10:50.59610005 +0000 UTC m=+101.103603071" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.643936 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podStartSLOduration=74.643914555 podStartE2EDuration="1m14.643914555s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.626807448 +0000 UTC m=+101.134310489" watchObservedRunningTime="2025-11-25 12:10:50.643914555 +0000 UTC m=+101.151417576" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.644236 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-5fwzm" podStartSLOduration=74.644231524 podStartE2EDuration="1m14.644231524s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.643748651 +0000 UTC m=+101.151251702" watchObservedRunningTime="2025-11-25 12:10:50.644231524 +0000 UTC m=+101.151734545" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.656516 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.65649752 podStartE2EDuration="45.65649752s" podCreationTimestamp="2025-11-25 12:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.655923214 +0000 UTC m=+101.163426245" watchObservedRunningTime="2025-11-25 12:10:50.65649752 +0000 UTC m=+101.164000531" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.669339 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.669321269 podStartE2EDuration="10.669321269s" podCreationTimestamp="2025-11-25 12:10:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:50.668287351 +0000 UTC m=+101.175790372" watchObservedRunningTime="2025-11-25 12:10:50.669321269 +0000 UTC m=+101.176824290" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.693703 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.693703 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.693826 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.693909 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:50 crc kubenswrapper[4715]: E1125 12:10:50.693952 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:50 crc kubenswrapper[4715]: E1125 12:10:50.694279 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:50 crc kubenswrapper[4715]: E1125 12:10:50.694390 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:50 crc kubenswrapper[4715]: E1125 12:10:50.694571 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.694739 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:10:50 crc kubenswrapper[4715]: E1125 12:10:50.694922 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" Nov 25 12:10:50 crc kubenswrapper[4715]: I1125 12:10:50.741478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" Nov 25 12:10:51 crc kubenswrapper[4715]: I1125 12:10:51.332684 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" event={"ID":"f145f4d6-2e9e-4d9e-afb0-fe334f747b27","Type":"ContainerStarted","Data":"4d8b91e51ae4ef7fb294944031a63476370bca1d907ba646e536c28fd36e4837"} Nov 25 12:10:51 crc kubenswrapper[4715]: I1125 12:10:51.333016 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" event={"ID":"f145f4d6-2e9e-4d9e-afb0-fe334f747b27","Type":"ContainerStarted","Data":"842a600fe5e97e82ab0e676979e0166bacf7d8a7b88c644313c4205a1e1abc82"} Nov 25 12:10:51 crc kubenswrapper[4715]: I1125 12:10:51.345220 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8tv2p" podStartSLOduration=75.345198521 podStartE2EDuration="1m15.345198521s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:10:51.344017608 +0000 UTC m=+101.851520639" watchObservedRunningTime="2025-11-25 12:10:51.345198521 +0000 UTC m=+101.852701542" Nov 25 12:10:52 crc kubenswrapper[4715]: I1125 12:10:52.692975 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:52 crc kubenswrapper[4715]: I1125 12:10:52.693048 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:52 crc kubenswrapper[4715]: I1125 12:10:52.693132 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:52 crc kubenswrapper[4715]: E1125 12:10:52.693269 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:52 crc kubenswrapper[4715]: I1125 12:10:52.693347 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:52 crc kubenswrapper[4715]: E1125 12:10:52.693503 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:52 crc kubenswrapper[4715]: E1125 12:10:52.693537 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:52 crc kubenswrapper[4715]: E1125 12:10:52.693597 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:54 crc kubenswrapper[4715]: I1125 12:10:54.693277 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:54 crc kubenswrapper[4715]: I1125 12:10:54.693429 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:54 crc kubenswrapper[4715]: I1125 12:10:54.693302 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:54 crc kubenswrapper[4715]: E1125 12:10:54.693588 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:54 crc kubenswrapper[4715]: E1125 12:10:54.693447 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:54 crc kubenswrapper[4715]: I1125 12:10:54.693277 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:54 crc kubenswrapper[4715]: E1125 12:10:54.693719 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:54 crc kubenswrapper[4715]: E1125 12:10:54.693786 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:55 crc kubenswrapper[4715]: I1125 12:10:55.217405 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:55 crc kubenswrapper[4715]: E1125 12:10:55.217603 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:10:55 crc kubenswrapper[4715]: E1125 12:10:55.217701 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs podName:3874a0dc-1a53-4587-bb14-7374a9715ae5 nodeName:}" failed. No retries permitted until 2025-11-25 12:11:59.217679582 +0000 UTC m=+169.725182603 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs") pod "network-metrics-daemon-c8s6d" (UID: "3874a0dc-1a53-4587-bb14-7374a9715ae5") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 12:10:56 crc kubenswrapper[4715]: I1125 12:10:56.693651 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:56 crc kubenswrapper[4715]: I1125 12:10:56.693811 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:56 crc kubenswrapper[4715]: I1125 12:10:56.693651 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:56 crc kubenswrapper[4715]: E1125 12:10:56.693887 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:56 crc kubenswrapper[4715]: I1125 12:10:56.693965 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:56 crc kubenswrapper[4715]: E1125 12:10:56.694137 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:56 crc kubenswrapper[4715]: E1125 12:10:56.694364 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:10:56 crc kubenswrapper[4715]: E1125 12:10:56.694518 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:58 crc kubenswrapper[4715]: I1125 12:10:58.692979 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:10:58 crc kubenswrapper[4715]: I1125 12:10:58.693028 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:10:58 crc kubenswrapper[4715]: I1125 12:10:58.693078 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:10:58 crc kubenswrapper[4715]: I1125 12:10:58.693016 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:10:58 crc kubenswrapper[4715]: E1125 12:10:58.693132 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:10:58 crc kubenswrapper[4715]: E1125 12:10:58.693230 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:10:58 crc kubenswrapper[4715]: E1125 12:10:58.693291 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:10:58 crc kubenswrapper[4715]: E1125 12:10:58.693339 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:00 crc kubenswrapper[4715]: I1125 12:11:00.692869 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:00 crc kubenswrapper[4715]: E1125 12:11:00.693969 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:00 crc kubenswrapper[4715]: I1125 12:11:00.693986 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:00 crc kubenswrapper[4715]: E1125 12:11:00.694133 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:00 crc kubenswrapper[4715]: I1125 12:11:00.694045 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:00 crc kubenswrapper[4715]: E1125 12:11:00.694225 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:00 crc kubenswrapper[4715]: I1125 12:11:00.694025 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:00 crc kubenswrapper[4715]: E1125 12:11:00.694292 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:02 crc kubenswrapper[4715]: I1125 12:11:02.694079 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:02 crc kubenswrapper[4715]: I1125 12:11:02.694098 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:02 crc kubenswrapper[4715]: E1125 12:11:02.694301 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:02 crc kubenswrapper[4715]: E1125 12:11:02.694402 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:02 crc kubenswrapper[4715]: I1125 12:11:02.694765 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:02 crc kubenswrapper[4715]: I1125 12:11:02.694850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:02 crc kubenswrapper[4715]: E1125 12:11:02.694961 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:02 crc kubenswrapper[4715]: E1125 12:11:02.695034 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:02 crc kubenswrapper[4715]: I1125 12:11:02.695238 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:11:02 crc kubenswrapper[4715]: E1125 12:11:02.695488 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" Nov 25 12:11:04 crc kubenswrapper[4715]: I1125 12:11:04.693555 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:04 crc kubenswrapper[4715]: I1125 12:11:04.693648 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:04 crc kubenswrapper[4715]: I1125 12:11:04.693805 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:04 crc kubenswrapper[4715]: I1125 12:11:04.693827 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:04 crc kubenswrapper[4715]: E1125 12:11:04.693807 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:04 crc kubenswrapper[4715]: E1125 12:11:04.693910 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:04 crc kubenswrapper[4715]: E1125 12:11:04.694075 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:04 crc kubenswrapper[4715]: E1125 12:11:04.694143 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:06 crc kubenswrapper[4715]: I1125 12:11:06.693876 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:06 crc kubenswrapper[4715]: I1125 12:11:06.693918 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:06 crc kubenswrapper[4715]: I1125 12:11:06.693974 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:06 crc kubenswrapper[4715]: E1125 12:11:06.694024 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:06 crc kubenswrapper[4715]: E1125 12:11:06.694141 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:06 crc kubenswrapper[4715]: I1125 12:11:06.694221 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:06 crc kubenswrapper[4715]: E1125 12:11:06.694273 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:06 crc kubenswrapper[4715]: E1125 12:11:06.694325 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:08 crc kubenswrapper[4715]: I1125 12:11:08.694024 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:08 crc kubenswrapper[4715]: I1125 12:11:08.694133 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:08 crc kubenswrapper[4715]: I1125 12:11:08.694207 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:08 crc kubenswrapper[4715]: E1125 12:11:08.694238 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:08 crc kubenswrapper[4715]: I1125 12:11:08.694050 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:08 crc kubenswrapper[4715]: E1125 12:11:08.694329 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:08 crc kubenswrapper[4715]: E1125 12:11:08.694395 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:08 crc kubenswrapper[4715]: E1125 12:11:08.694437 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:10 crc kubenswrapper[4715]: E1125 12:11:10.665004 4715 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 25 12:11:10 crc kubenswrapper[4715]: I1125 12:11:10.693060 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:10 crc kubenswrapper[4715]: I1125 12:11:10.693096 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:10 crc kubenswrapper[4715]: I1125 12:11:10.693060 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:10 crc kubenswrapper[4715]: E1125 12:11:10.693219 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:10 crc kubenswrapper[4715]: I1125 12:11:10.693290 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:10 crc kubenswrapper[4715]: E1125 12:11:10.694527 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:10 crc kubenswrapper[4715]: E1125 12:11:10.694730 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:10 crc kubenswrapper[4715]: E1125 12:11:10.694870 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:10 crc kubenswrapper[4715]: E1125 12:11:10.987242 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 12:11:12 crc kubenswrapper[4715]: I1125 12:11:12.693871 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:12 crc kubenswrapper[4715]: I1125 12:11:12.693941 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:12 crc kubenswrapper[4715]: E1125 12:11:12.694109 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:12 crc kubenswrapper[4715]: I1125 12:11:12.694456 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:12 crc kubenswrapper[4715]: E1125 12:11:12.694527 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:12 crc kubenswrapper[4715]: I1125 12:11:12.694579 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:12 crc kubenswrapper[4715]: E1125 12:11:12.694715 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:12 crc kubenswrapper[4715]: E1125 12:11:12.694816 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:13 crc kubenswrapper[4715]: I1125 12:11:13.693985 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:11:13 crc kubenswrapper[4715]: E1125 12:11:13.694232 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tzl95_openshift-ovn-kubernetes(02385935-e23a-4708-b57f-fd393d05651b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.410318 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/1.log" Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.410924 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/0.log" Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.410996 4715 generic.go:334] "Generic (PLEG): container finished" podID="6371ac18-2361-43bb-b474-32bf53713cf5" containerID="2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d" exitCode=1 Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.411043 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5mzr5" event={"ID":"6371ac18-2361-43bb-b474-32bf53713cf5","Type":"ContainerDied","Data":"2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d"} Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.411105 4715 scope.go:117] "RemoveContainer" containerID="eb94e657c200ae2afcca89c3c37e75f8333efe103dac43bb6f62ede793e5c798" Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.411664 4715 scope.go:117] "RemoveContainer" containerID="2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d" Nov 25 12:11:14 crc kubenswrapper[4715]: E1125 12:11:14.411873 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-5mzr5_openshift-multus(6371ac18-2361-43bb-b474-32bf53713cf5)\"" pod="openshift-multus/multus-5mzr5" podUID="6371ac18-2361-43bb-b474-32bf53713cf5" Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.693015 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.693094 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.693102 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:14 crc kubenswrapper[4715]: E1125 12:11:14.693868 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:14 crc kubenswrapper[4715]: E1125 12:11:14.693868 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:14 crc kubenswrapper[4715]: I1125 12:11:14.693105 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:14 crc kubenswrapper[4715]: E1125 12:11:14.694280 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:14 crc kubenswrapper[4715]: E1125 12:11:14.694026 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:15 crc kubenswrapper[4715]: I1125 12:11:15.416708 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/1.log" Nov 25 12:11:15 crc kubenswrapper[4715]: E1125 12:11:15.989101 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 12:11:16 crc kubenswrapper[4715]: I1125 12:11:16.693754 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:16 crc kubenswrapper[4715]: I1125 12:11:16.693820 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:16 crc kubenswrapper[4715]: I1125 12:11:16.693870 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:16 crc kubenswrapper[4715]: E1125 12:11:16.694839 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:16 crc kubenswrapper[4715]: I1125 12:11:16.694740 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:16 crc kubenswrapper[4715]: E1125 12:11:16.694956 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:16 crc kubenswrapper[4715]: E1125 12:11:16.694584 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:16 crc kubenswrapper[4715]: E1125 12:11:16.694854 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:18 crc kubenswrapper[4715]: I1125 12:11:18.692938 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:18 crc kubenswrapper[4715]: I1125 12:11:18.693014 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:18 crc kubenswrapper[4715]: I1125 12:11:18.692965 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:18 crc kubenswrapper[4715]: E1125 12:11:18.693133 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:18 crc kubenswrapper[4715]: I1125 12:11:18.693145 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:18 crc kubenswrapper[4715]: E1125 12:11:18.693279 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:18 crc kubenswrapper[4715]: E1125 12:11:18.693378 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:18 crc kubenswrapper[4715]: E1125 12:11:18.693411 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:20 crc kubenswrapper[4715]: I1125 12:11:20.693293 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:20 crc kubenswrapper[4715]: I1125 12:11:20.693347 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:20 crc kubenswrapper[4715]: E1125 12:11:20.694651 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:20 crc kubenswrapper[4715]: I1125 12:11:20.694686 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:20 crc kubenswrapper[4715]: I1125 12:11:20.694668 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:20 crc kubenswrapper[4715]: E1125 12:11:20.694757 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:20 crc kubenswrapper[4715]: E1125 12:11:20.694823 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:20 crc kubenswrapper[4715]: E1125 12:11:20.694865 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:20 crc kubenswrapper[4715]: E1125 12:11:20.989961 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 12:11:22 crc kubenswrapper[4715]: I1125 12:11:22.692988 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:22 crc kubenswrapper[4715]: I1125 12:11:22.692988 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:22 crc kubenswrapper[4715]: I1125 12:11:22.693024 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:22 crc kubenswrapper[4715]: E1125 12:11:22.693224 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:22 crc kubenswrapper[4715]: E1125 12:11:22.693317 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:22 crc kubenswrapper[4715]: I1125 12:11:22.693365 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:22 crc kubenswrapper[4715]: E1125 12:11:22.693435 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:22 crc kubenswrapper[4715]: E1125 12:11:22.693495 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:24 crc kubenswrapper[4715]: I1125 12:11:24.693737 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:24 crc kubenswrapper[4715]: I1125 12:11:24.693792 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:24 crc kubenswrapper[4715]: I1125 12:11:24.693810 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:24 crc kubenswrapper[4715]: I1125 12:11:24.693765 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:24 crc kubenswrapper[4715]: E1125 12:11:24.693999 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:24 crc kubenswrapper[4715]: E1125 12:11:24.693886 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:24 crc kubenswrapper[4715]: E1125 12:11:24.694117 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:24 crc kubenswrapper[4715]: E1125 12:11:24.694180 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:25 crc kubenswrapper[4715]: E1125 12:11:25.991147 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 12:11:26 crc kubenswrapper[4715]: I1125 12:11:26.693741 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:26 crc kubenswrapper[4715]: I1125 12:11:26.693779 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:26 crc kubenswrapper[4715]: I1125 12:11:26.693887 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:26 crc kubenswrapper[4715]: E1125 12:11:26.693881 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:26 crc kubenswrapper[4715]: I1125 12:11:26.693996 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:26 crc kubenswrapper[4715]: E1125 12:11:26.694074 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:26 crc kubenswrapper[4715]: E1125 12:11:26.694230 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:26 crc kubenswrapper[4715]: E1125 12:11:26.694278 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:26 crc kubenswrapper[4715]: I1125 12:11:26.694954 4715 scope.go:117] "RemoveContainer" containerID="2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d" Nov 25 12:11:26 crc kubenswrapper[4715]: I1125 12:11:26.695008 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:11:27 crc kubenswrapper[4715]: I1125 12:11:27.460623 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/1.log" Nov 25 12:11:27 crc kubenswrapper[4715]: I1125 12:11:27.460926 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5mzr5" event={"ID":"6371ac18-2361-43bb-b474-32bf53713cf5","Type":"ContainerStarted","Data":"3579f1d8f85e742d8f3051d9ab2cc486b0cd4fdc99124c1454343e4bda4a8925"} Nov 25 12:11:27 crc kubenswrapper[4715]: I1125 12:11:27.463173 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/3.log" Nov 25 12:11:27 crc kubenswrapper[4715]: I1125 12:11:27.465980 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerStarted","Data":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} Nov 25 12:11:27 crc kubenswrapper[4715]: I1125 12:11:27.466388 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:11:27 crc kubenswrapper[4715]: I1125 12:11:27.504456 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podStartSLOduration=111.504437876 podStartE2EDuration="1m51.504437876s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:27.503705856 +0000 UTC m=+138.011208877" watchObservedRunningTime="2025-11-25 12:11:27.504437876 +0000 UTC m=+138.011940897" Nov 25 12:11:27 crc kubenswrapper[4715]: I1125 12:11:27.713666 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-c8s6d"] Nov 25 12:11:27 crc kubenswrapper[4715]: I1125 12:11:27.713807 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:27 crc kubenswrapper[4715]: E1125 12:11:27.713906 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:28 crc kubenswrapper[4715]: I1125 12:11:28.693831 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:28 crc kubenswrapper[4715]: E1125 12:11:28.693965 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:28 crc kubenswrapper[4715]: I1125 12:11:28.694013 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:28 crc kubenswrapper[4715]: I1125 12:11:28.694264 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:28 crc kubenswrapper[4715]: E1125 12:11:28.694363 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:28 crc kubenswrapper[4715]: E1125 12:11:28.694401 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:29 crc kubenswrapper[4715]: I1125 12:11:29.693343 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:29 crc kubenswrapper[4715]: E1125 12:11:29.693490 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-c8s6d" podUID="3874a0dc-1a53-4587-bb14-7374a9715ae5" Nov 25 12:11:30 crc kubenswrapper[4715]: I1125 12:11:30.694023 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:30 crc kubenswrapper[4715]: I1125 12:11:30.694030 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:30 crc kubenswrapper[4715]: I1125 12:11:30.694072 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:30 crc kubenswrapper[4715]: E1125 12:11:30.694545 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 12:11:30 crc kubenswrapper[4715]: E1125 12:11:30.694665 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 12:11:30 crc kubenswrapper[4715]: E1125 12:11:30.694776 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 12:11:31 crc kubenswrapper[4715]: I1125 12:11:31.693741 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:31 crc kubenswrapper[4715]: I1125 12:11:31.696722 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 12:11:31 crc kubenswrapper[4715]: I1125 12:11:31.696730 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 12:11:32 crc kubenswrapper[4715]: I1125 12:11:32.693647 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:32 crc kubenswrapper[4715]: I1125 12:11:32.693686 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:32 crc kubenswrapper[4715]: I1125 12:11:32.693702 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:32 crc kubenswrapper[4715]: I1125 12:11:32.696951 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 12:11:32 crc kubenswrapper[4715]: I1125 12:11:32.697032 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 12:11:32 crc kubenswrapper[4715]: I1125 12:11:32.697251 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 12:11:32 crc kubenswrapper[4715]: I1125 12:11:32.697483 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 12:11:33 crc kubenswrapper[4715]: I1125 12:11:33.729482 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.597746 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.598692 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.699213 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.699417 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:40 crc kubenswrapper[4715]: E1125 12:11:40.699551 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:13:42.69949781 +0000 UTC m=+273.207000931 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.699683 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.699840 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.705810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.705913 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.706535 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.808838 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.816930 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 12:11:40 crc kubenswrapper[4715]: I1125 12:11:40.823278 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.043387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.076589 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.077259 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.077657 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dl4rf"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.078382 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.081548 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7vcvm"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.081989 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4blqh"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.082388 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.082669 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.083431 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.084501 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.085627 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.086017 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.099525 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.099856 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.100283 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.100469 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.100550 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.100574 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.100635 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.100703 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.100960 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101082 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101140 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101086 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101270 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101330 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101360 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101405 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101445 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101465 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101238 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101592 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.101930 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.102053 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.102253 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.102504 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.102674 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.102769 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.102957 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.103051 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.103142 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.103230 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jtjqf"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.103398 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.103546 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.103656 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.103773 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.103801 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.103945 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.104055 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.104513 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.108818 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.111612 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.112387 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.112632 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.113260 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.116997 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.117190 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.117554 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.117643 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.117686 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.117720 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tw7jp"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.117733 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.117806 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.117560 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.118170 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.118287 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.118405 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5clnl"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.118676 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.118935 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.119142 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.119643 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.119735 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qztrg"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.119773 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.120075 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.124473 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.140816 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.141696 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.141982 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-956n7"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.143224 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.143319 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.143489 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.143838 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.150713 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.164217 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-4dfdh"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.164832 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.165083 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-tvx5q"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.165486 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.165524 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tvx5q" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.165613 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.165893 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.165971 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166012 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166041 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166273 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166444 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166687 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166707 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166736 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166800 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166695 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166875 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166932 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166930 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166937 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166966 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166980 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.166872 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.167697 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zrx9b"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.168225 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.168762 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kbdhr"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.169296 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.170233 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.170430 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.170596 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.170703 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.170801 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.170903 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.170918 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.171011 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.171096 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.171216 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.171309 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.171408 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.171528 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.171651 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.173231 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.173432 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.173591 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.173687 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.174791 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.175301 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.175345 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.175388 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.175454 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.175658 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.175844 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.176458 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.176498 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.176584 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.180820 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.181560 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.181097 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.181748 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.184459 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-f6n6v"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.194970 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.196757 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.197617 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.198038 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.198435 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.199423 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rstj"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.199872 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.200261 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.202754 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.202940 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.203729 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.202262 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.202318 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.230058 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.231043 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-vn6lh"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.233054 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7vcvm"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.233080 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dl4rf"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.233311 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.246086 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.246386 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.247449 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.248642 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.248883 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-bound-sa-token\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.248935 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtdwp\" (UniqueName: \"kubernetes.io/projected/3d62103e-bb94-453d-b9cb-9a0de58823ce-kube-api-access-dtdwp\") pod \"openshift-controller-manager-operator-756b6f6bc6-fwzk5\" (UID: \"3d62103e-bb94-453d-b9cb-9a0de58823ce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.248968 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f54afd7f-6b85-41b7-9a24-c253e16b8d88-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.248990 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.248998 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249011 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-client-ca\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249036 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f54afd7f-6b85-41b7-9a24-c253e16b8d88-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x45kf\" (UniqueName: \"kubernetes.io/projected/4ba7f02a-888d-4def-a53b-7a39407051fb-kube-api-access-x45kf\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249087 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249105 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/209069f6-4788-4ce3-ac7f-6446f4c26e05-etcd-client\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249127 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/209069f6-4788-4ce3-ac7f-6446f4c26e05-audit-dir\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249148 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a3f84e-2d87-402d-8f01-c4965cd69dee-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wmpt2\" (UID: \"b2a3f84e-2d87-402d-8f01-c4965cd69dee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249186 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmkz6\" (UniqueName: \"kubernetes.io/projected/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-kube-api-access-gmkz6\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249238 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gznfl\" (UniqueName: \"kubernetes.io/projected/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-kube-api-access-gznfl\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249270 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/39830646-31e2-4717-9497-d499e89d16c3-auth-proxy-config\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249308 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f37a4a33-7308-4546-a042-6b7335fbeccf-serving-cert\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249333 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5cf156ab-e8de-4545-b432-dd538b3c7ce4-etcd-client\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249389 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5f6v\" (UniqueName: \"kubernetes.io/projected/5cf156ab-e8de-4545-b432-dd538b3c7ce4-kube-api-access-n5f6v\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249421 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46btr\" (UniqueName: \"kubernetes.io/projected/209069f6-4788-4ce3-ac7f-6446f4c26e05-kube-api-access-46btr\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249451 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249471 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-etcd-serving-ca\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249490 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-metrics-tls\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249513 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4528d803-2051-4abb-9af2-898361f7b2fc-serving-cert\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249587 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f54afd7f-6b85-41b7-9a24-c253e16b8d88-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249609 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249629 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249652 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249846 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-trusted-ca\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/209069f6-4788-4ce3-ac7f-6446f4c26e05-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249902 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f37a4a33-7308-4546-a042-6b7335fbeccf-config\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249931 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2374ca5a-3a6e-461b-9d9d-84d95a37a804-images\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249960 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2374ca5a-3a6e-461b-9d9d-84d95a37a804-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.249982 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d62103e-bb94-453d-b9cb-9a0de58823ce-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fwzk5\" (UID: \"3d62103e-bb94-453d-b9cb-9a0de58823ce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250005 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-image-import-ca\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250028 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f37a4a33-7308-4546-a042-6b7335fbeccf-trusted-ca\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250055 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2374ca5a-3a6e-461b-9d9d-84d95a37a804-config\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250080 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250103 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-serving-cert\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250130 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d62103e-bb94-453d-b9cb-9a0de58823ce-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fwzk5\" (UID: \"3d62103e-bb94-453d-b9cb-9a0de58823ce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250188 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250290 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b5c80b59-9196-4573-ac01-eb88a738fa25-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qmkr2\" (UID: \"b5c80b59-9196-4573-ac01-eb88a738fa25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250313 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cmfq\" (UniqueName: \"kubernetes.io/projected/f40f7173-969d-45d1-baaa-8ed0eaab8499-kube-api-access-9cmfq\") pod \"openshift-config-operator-7777fb866f-cj9n6\" (UID: \"f40f7173-969d-45d1-baaa-8ed0eaab8499\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250342 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba7f02a-888d-4def-a53b-7a39407051fb-serving-cert\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250365 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-config\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250389 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250415 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2a3f84e-2d87-402d-8f01-c4965cd69dee-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wmpt2\" (UID: \"b2a3f84e-2d87-402d-8f01-c4965cd69dee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250438 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cf156ab-e8de-4545-b432-dd538b3c7ce4-node-pullsecrets\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250464 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/78c80454-cff3-4429-aafe-4359423f9b68-metrics-tls\") pod \"dns-operator-744455d44c-tw7jp\" (UID: \"78c80454-cff3-4429-aafe-4359423f9b68\") " pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250511 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-policies\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250557 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-dir\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250596 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/209069f6-4788-4ce3-ac7f-6446f4c26e05-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250646 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-config\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250668 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250693 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-client-ca\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250722 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvtc6\" (UniqueName: \"kubernetes.io/projected/4528d803-2051-4abb-9af2-898361f7b2fc-kube-api-access-hvtc6\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250757 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f40f7173-969d-45d1-baaa-8ed0eaab8499-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cj9n6\" (UID: \"f40f7173-969d-45d1-baaa-8ed0eaab8499\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250785 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cf156ab-e8de-4545-b432-dd538b3c7ce4-serving-cert\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250813 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250833 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/39830646-31e2-4717-9497-d499e89d16c3-machine-approver-tls\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250852 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39830646-31e2-4717-9497-d499e89d16c3-config\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250874 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-config\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn9tj\" (UniqueName: \"kubernetes.io/projected/78c80454-cff3-4429-aafe-4359423f9b68-kube-api-access-sn9tj\") pod \"dns-operator-744455d44c-tw7jp\" (UID: \"78c80454-cff3-4429-aafe-4359423f9b68\") " pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250924 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6kl5\" (UniqueName: \"kubernetes.io/projected/39830646-31e2-4717-9497-d499e89d16c3-kube-api-access-p6kl5\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250949 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250972 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5cf156ab-e8de-4545-b432-dd538b3c7ce4-audit-dir\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.250991 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-service-ca-bundle\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251011 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/209069f6-4788-4ce3-ac7f-6446f4c26e05-serving-cert\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251031 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r2nn\" (UniqueName: \"kubernetes.io/projected/f37a4a33-7308-4546-a042-6b7335fbeccf-kube-api-access-2r2nn\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251050 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f40f7173-969d-45d1-baaa-8ed0eaab8499-serving-cert\") pod \"openshift-config-operator-7777fb866f-cj9n6\" (UID: \"f40f7173-969d-45d1-baaa-8ed0eaab8499\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251070 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/209069f6-4788-4ce3-ac7f-6446f4c26e05-audit-policies\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251094 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9fpp\" (UniqueName: \"kubernetes.io/projected/2374ca5a-3a6e-461b-9d9d-84d95a37a804-kube-api-access-d9fpp\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251114 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n46p9\" (UniqueName: \"kubernetes.io/projected/b5c80b59-9196-4573-ac01-eb88a738fa25-kube-api-access-n46p9\") pod \"cluster-samples-operator-665b6dd947-qmkr2\" (UID: \"b5c80b59-9196-4573-ac01-eb88a738fa25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251134 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nttg6\" (UniqueName: \"kubernetes.io/projected/b2a3f84e-2d87-402d-8f01-c4965cd69dee-kube-api-access-nttg6\") pod \"openshift-apiserver-operator-796bbdcf4f-wmpt2\" (UID: \"b2a3f84e-2d87-402d-8f01-c4965cd69dee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251159 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251186 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-audit\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251226 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5cf156ab-e8de-4545-b432-dd538b3c7ce4-encryption-config\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251247 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-config\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251269 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhkrj\" (UniqueName: \"kubernetes.io/projected/f54afd7f-6b85-41b7-9a24-c253e16b8d88-kube-api-access-xhkrj\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251293 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/209069f6-4788-4ce3-ac7f-6446f4c26e05-encryption-config\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251314 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls47w\" (UniqueName: \"kubernetes.io/projected/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-kube-api-access-ls47w\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.251970 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.252779 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.253522 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.259856 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.263458 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.263516 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.268800 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.268912 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.269124 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.269272 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.269528 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wslh5"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.270097 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.270245 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.270407 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.282037 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.282770 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.284971 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.285329 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.286068 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4blqh"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.286118 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.286135 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2xgrz"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.286337 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.286540 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.287342 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.287370 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.287383 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jtjqf"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.287451 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.289480 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.289988 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.292295 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.295308 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.295372 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.295386 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rstj"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.298572 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-956n7"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.298648 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5clnl"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.301572 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.301649 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-f6n6v"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.312367 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.312537 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.313498 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tvx5q"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.318064 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tw7jp"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.320384 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.323687 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4dfdh"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.324426 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kbdhr"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.329413 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.331634 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.342519 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wslh5"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.345172 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.347815 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.348083 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.352021 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.352976 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d62103e-bb94-453d-b9cb-9a0de58823ce-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fwzk5\" (UID: \"3d62103e-bb94-453d-b9cb-9a0de58823ce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353029 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fa028433-d9fc-4729-bdb0-65cd9a4695e5-images\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353050 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353071 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b5c80b59-9196-4573-ac01-eb88a738fa25-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qmkr2\" (UID: \"b5c80b59-9196-4573-ac01-eb88a738fa25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353090 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1e0bb7e2-0204-4e78-afb5-6d0ecfd59541-profile-collector-cert\") pod \"catalog-operator-68c6474976-5grcc\" (UID: \"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353108 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b392c0-15c6-436c-9406-c410413c8f9d-config\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353124 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/26b392c0-15c6-436c-9406-c410413c8f9d-etcd-client\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353141 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-config\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353160 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z4p9\" (UniqueName: \"kubernetes.io/projected/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-kube-api-access-7z4p9\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353177 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2a3f84e-2d87-402d-8f01-c4965cd69dee-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wmpt2\" (UID: \"b2a3f84e-2d87-402d-8f01-c4965cd69dee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353225 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cf156ab-e8de-4545-b432-dd538b3c7ce4-node-pullsecrets\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353242 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-dir\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353256 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/209069f6-4788-4ce3-ac7f-6446f4c26e05-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353274 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05d83c7d-74cd-4b0b-871e-4f522ed1775d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-htvkk\" (UID: \"05d83c7d-74cd-4b0b-871e-4f522ed1775d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353294 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4z6t\" (UniqueName: \"kubernetes.io/projected/19cbbcfc-8187-4b59-a1e3-3a33e64e37c4-kube-api-access-p4z6t\") pod \"multus-admission-controller-857f4d67dd-f6n6v\" (UID: \"19cbbcfc-8187-4b59-a1e3-3a33e64e37c4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353314 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f40f7173-969d-45d1-baaa-8ed0eaab8499-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cj9n6\" (UID: \"f40f7173-969d-45d1-baaa-8ed0eaab8499\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353334 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cf156ab-e8de-4545-b432-dd538b3c7ce4-serving-cert\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353350 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353368 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/39830646-31e2-4717-9497-d499e89d16c3-machine-approver-tls\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39830646-31e2-4717-9497-d499e89d16c3-config\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353397 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-config\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353414 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn9tj\" (UniqueName: \"kubernetes.io/projected/78c80454-cff3-4429-aafe-4359423f9b68-kube-api-access-sn9tj\") pod \"dns-operator-744455d44c-tw7jp\" (UID: \"78c80454-cff3-4429-aafe-4359423f9b68\") " pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353430 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2353d56f-1cab-4181-b1c2-164a90efab5e-service-ca-bundle\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353447 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6kl5\" (UniqueName: \"kubernetes.io/projected/39830646-31e2-4717-9497-d499e89d16c3-kube-api-access-p6kl5\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353464 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa028433-d9fc-4729-bdb0-65cd9a4695e5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353482 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfbxx\" (UniqueName: \"kubernetes.io/projected/1e0bb7e2-0204-4e78-afb5-6d0ecfd59541-kube-api-access-rfbxx\") pod \"catalog-operator-68c6474976-5grcc\" (UID: \"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353504 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353531 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r2nn\" (UniqueName: \"kubernetes.io/projected/f37a4a33-7308-4546-a042-6b7335fbeccf-kube-api-access-2r2nn\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353556 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00cb67a6-0595-4968-a2d5-3a420bbd3c09-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9h2w\" (UID: \"00cb67a6-0595-4968-a2d5-3a420bbd3c09\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353580 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/209069f6-4788-4ce3-ac7f-6446f4c26e05-serving-cert\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353592 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-dir\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353711 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cf156ab-e8de-4545-b432-dd538b3c7ce4-node-pullsecrets\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353602 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/209069f6-4788-4ce3-ac7f-6446f4c26e05-audit-policies\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353753 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2353d56f-1cab-4181-b1c2-164a90efab5e-metrics-certs\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353781 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-audit\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353823 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5cf156ab-e8de-4545-b432-dd538b3c7ce4-encryption-config\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353844 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-config\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353872 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kqt2\" (UniqueName: \"kubernetes.io/projected/fa028433-d9fc-4729-bdb0-65cd9a4695e5-kube-api-access-4kqt2\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353901 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls47w\" (UniqueName: \"kubernetes.io/projected/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-kube-api-access-ls47w\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353908 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d62103e-bb94-453d-b9cb-9a0de58823ce-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fwzk5\" (UID: \"3d62103e-bb94-453d-b9cb-9a0de58823ce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.353928 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-bound-sa-token\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.354026 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtdwp\" (UniqueName: \"kubernetes.io/projected/3d62103e-bb94-453d-b9cb-9a0de58823ce-kube-api-access-dtdwp\") pod \"openshift-controller-manager-operator-756b6f6bc6-fwzk5\" (UID: \"3d62103e-bb94-453d-b9cb-9a0de58823ce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.354049 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f54afd7f-6b85-41b7-9a24-c253e16b8d88-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.354068 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.354083 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-client-ca\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.354105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/209069f6-4788-4ce3-ac7f-6446f4c26e05-audit-dir\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.354123 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5wpt\" (UniqueName: \"kubernetes.io/projected/00cb67a6-0595-4968-a2d5-3a420bbd3c09-kube-api-access-m5wpt\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9h2w\" (UID: \"00cb67a6-0595-4968-a2d5-3a420bbd3c09\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.354774 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-audit\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355123 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/209069f6-4788-4ce3-ac7f-6446f4c26e05-audit-dir\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355209 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-oauth-serving-cert\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355274 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gznfl\" (UniqueName: \"kubernetes.io/projected/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-kube-api-access-gznfl\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355301 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/805eb165-93ad-4085-92bc-28ad037ed23c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jx2z4\" (UID: \"805eb165-93ad-4085-92bc-28ad037ed23c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355356 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f37a4a33-7308-4546-a042-6b7335fbeccf-serving-cert\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355383 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w75s\" (UniqueName: \"kubernetes.io/projected/13c92ba7-a34e-4f49-b98d-032d6d8543db-kube-api-access-5w75s\") pod \"downloads-7954f5f757-tvx5q\" (UID: \"13c92ba7-a34e-4f49-b98d-032d6d8543db\") " pod="openshift-console/downloads-7954f5f757-tvx5q" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355410 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2rstj\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355435 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46btr\" (UniqueName: \"kubernetes.io/projected/209069f6-4788-4ce3-ac7f-6446f4c26e05-kube-api-access-46btr\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355503 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/05d83c7d-74cd-4b0b-871e-4f522ed1775d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-htvkk\" (UID: \"05d83c7d-74cd-4b0b-871e-4f522ed1775d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355528 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-metrics-tls\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355561 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355586 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05d83c7d-74cd-4b0b-871e-4f522ed1775d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-htvkk\" (UID: \"05d83c7d-74cd-4b0b-871e-4f522ed1775d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355613 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f54afd7f-6b85-41b7-9a24-c253e16b8d88-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355690 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2353d56f-1cab-4181-b1c2-164a90efab5e-default-certificate\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355715 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f37a4a33-7308-4546-a042-6b7335fbeccf-config\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-image-import-ca\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355757 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f37a4a33-7308-4546-a042-6b7335fbeccf-trusted-ca\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355779 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-console-config\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355801 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m572w\" (UniqueName: \"kubernetes.io/projected/6467e871-dd98-4bcf-bf7c-53726589346b-kube-api-access-m572w\") pod \"marketplace-operator-79b997595-2rstj\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355826 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2374ca5a-3a6e-461b-9d9d-84d95a37a804-images\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355853 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2374ca5a-3a6e-461b-9d9d-84d95a37a804-config\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.355879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-serving-cert\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.356622 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-config\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.356726 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.357049 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39830646-31e2-4717-9497-d499e89d16c3-config\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.357476 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f40f7173-969d-45d1-baaa-8ed0eaab8499-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cj9n6\" (UID: \"f40f7173-969d-45d1-baaa-8ed0eaab8499\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.358808 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2374ca5a-3a6e-461b-9d9d-84d95a37a804-images\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.359469 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-client-ca\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.359679 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-config\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.360399 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/209069f6-4788-4ce3-ac7f-6446f4c26e05-audit-policies\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.360499 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.360512 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/209069f6-4788-4ce3-ac7f-6446f4c26e05-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.360607 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zrx9b"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.360647 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qztrg"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.361561 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.361597 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2a3f84e-2d87-402d-8f01-c4965cd69dee-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wmpt2\" (UID: \"b2a3f84e-2d87-402d-8f01-c4965cd69dee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.361758 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cmfq\" (UniqueName: \"kubernetes.io/projected/f40f7173-969d-45d1-baaa-8ed0eaab8499-kube-api-access-9cmfq\") pod \"openshift-config-operator-7777fb866f-cj9n6\" (UID: \"f40f7173-969d-45d1-baaa-8ed0eaab8499\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.361845 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26b392c0-15c6-436c-9406-c410413c8f9d-serving-cert\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.361864 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/209069f6-4788-4ce3-ac7f-6446f4c26e05-serving-cert\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.361912 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/39830646-31e2-4717-9497-d499e89d16c3-machine-approver-tls\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.361908 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-config\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.362308 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba7f02a-888d-4def-a53b-7a39407051fb-serving-cert\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.362352 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.362350 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b5c80b59-9196-4573-ac01-eb88a738fa25-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qmkr2\" (UID: \"b5c80b59-9196-4573-ac01-eb88a738fa25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.362519 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/78c80454-cff3-4429-aafe-4359423f9b68-metrics-tls\") pod \"dns-operator-744455d44c-tw7jp\" (UID: \"78c80454-cff3-4429-aafe-4359423f9b68\") " pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.362752 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/805eb165-93ad-4085-92bc-28ad037ed23c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jx2z4\" (UID: \"805eb165-93ad-4085-92bc-28ad037ed23c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.363429 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f37a4a33-7308-4546-a042-6b7335fbeccf-config\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.363692 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa028433-d9fc-4729-bdb0-65cd9a4695e5-proxy-tls\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.363768 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-policies\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.363798 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.363867 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-config\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364339 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f54afd7f-6b85-41b7-9a24-c253e16b8d88-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364360 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364475 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-client-ca\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364506 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5cf156ab-e8de-4545-b432-dd538b3c7ce4-encryption-config\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364510 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvtc6\" (UniqueName: \"kubernetes.io/projected/4528d803-2051-4abb-9af2-898361f7b2fc-kube-api-access-hvtc6\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364596 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-console-oauth-config\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364647 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cf156ab-e8de-4545-b432-dd538b3c7ce4-serving-cert\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2rstj\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364697 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5cf156ab-e8de-4545-b432-dd538b3c7ce4-audit-dir\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364725 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364849 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f37a4a33-7308-4546-a042-6b7335fbeccf-trusted-ca\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364861 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364895 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-service-ca-bundle\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364927 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/19cbbcfc-8187-4b59-a1e3-3a33e64e37c4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f6n6v\" (UID: \"19cbbcfc-8187-4b59-a1e3-3a33e64e37c4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364954 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f40f7173-969d-45d1-baaa-8ed0eaab8499-serving-cert\") pod \"openshift-config-operator-7777fb866f-cj9n6\" (UID: \"f40f7173-969d-45d1-baaa-8ed0eaab8499\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.364993 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9fpp\" (UniqueName: \"kubernetes.io/projected/2374ca5a-3a6e-461b-9d9d-84d95a37a804-kube-api-access-d9fpp\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365021 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n46p9\" (UniqueName: \"kubernetes.io/projected/b5c80b59-9196-4573-ac01-eb88a738fa25-kube-api-access-n46p9\") pod \"cluster-samples-operator-665b6dd947-qmkr2\" (UID: \"b5c80b59-9196-4573-ac01-eb88a738fa25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365119 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nttg6\" (UniqueName: \"kubernetes.io/projected/b2a3f84e-2d87-402d-8f01-c4965cd69dee-kube-api-access-nttg6\") pod \"openshift-apiserver-operator-796bbdcf4f-wmpt2\" (UID: \"b2a3f84e-2d87-402d-8f01-c4965cd69dee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365151 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/26b392c0-15c6-436c-9406-c410413c8f9d-etcd-service-ca\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365178 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgwc6\" (UniqueName: \"kubernetes.io/projected/26b392c0-15c6-436c-9406-c410413c8f9d-kube-api-access-rgwc6\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365233 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365260 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhkrj\" (UniqueName: \"kubernetes.io/projected/f54afd7f-6b85-41b7-9a24-c253e16b8d88-kube-api-access-xhkrj\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365284 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/209069f6-4788-4ce3-ac7f-6446f4c26e05-encryption-config\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365314 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxr6d\" (UniqueName: \"kubernetes.io/projected/2353d56f-1cab-4181-b1c2-164a90efab5e-kube-api-access-xxr6d\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365339 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/26b392c0-15c6-436c-9406-c410413c8f9d-etcd-ca\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365433 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365821 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2374ca5a-3a6e-461b-9d9d-84d95a37a804-config\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365963 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-metrics-tls\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366000 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-serving-cert\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366113 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-client-ca\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366396 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5cf156ab-e8de-4545-b432-dd538b3c7ce4-audit-dir\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366477 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-config\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.365289 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366573 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-service-ca-bundle\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366680 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp9b4\" (UniqueName: \"kubernetes.io/projected/946f1f86-8ba8-4e6c-975a-b185265eb557-kube-api-access-rp9b4\") pod \"migrator-59844c95c7-krnvf\" (UID: \"946f1f86-8ba8-4e6c-975a-b185265eb557\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366732 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2353d56f-1cab-4181-b1c2-164a90efab5e-stats-auth\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366809 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f54afd7f-6b85-41b7-9a24-c253e16b8d88-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366870 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x45kf\" (UniqueName: \"kubernetes.io/projected/4ba7f02a-888d-4def-a53b-7a39407051fb-kube-api-access-x45kf\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366900 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366924 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/209069f6-4788-4ce3-ac7f-6446f4c26e05-etcd-client\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.366948 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a3f84e-2d87-402d-8f01-c4965cd69dee-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wmpt2\" (UID: \"b2a3f84e-2d87-402d-8f01-c4965cd69dee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367259 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmkz6\" (UniqueName: \"kubernetes.io/projected/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-kube-api-access-gmkz6\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367433 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/39830646-31e2-4717-9497-d499e89d16c3-auth-proxy-config\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367466 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/805eb165-93ad-4085-92bc-28ad037ed23c-config\") pod \"kube-apiserver-operator-766d6c64bb-jx2z4\" (UID: \"805eb165-93ad-4085-92bc-28ad037ed23c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367491 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5cf156ab-e8de-4545-b432-dd538b3c7ce4-etcd-client\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367511 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5f6v\" (UniqueName: \"kubernetes.io/projected/5cf156ab-e8de-4545-b432-dd538b3c7ce4-kube-api-access-n5f6v\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367528 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-etcd-serving-ca\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367547 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367569 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4528d803-2051-4abb-9af2-898361f7b2fc-serving-cert\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367588 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-console-serving-cert\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367599 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a3f84e-2d87-402d-8f01-c4965cd69dee-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wmpt2\" (UID: \"b2a3f84e-2d87-402d-8f01-c4965cd69dee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367607 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cb67a6-0595-4968-a2d5-3a420bbd3c09-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9h2w\" (UID: \"00cb67a6-0595-4968-a2d5-3a420bbd3c09\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367661 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-trusted-ca\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367688 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-service-ca\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367711 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1e0bb7e2-0204-4e78-afb5-6d0ecfd59541-srv-cert\") pod \"catalog-operator-68c6474976-5grcc\" (UID: \"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367744 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367776 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367801 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/209069f6-4788-4ce3-ac7f-6446f4c26e05-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367832 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2374ca5a-3a6e-461b-9d9d-84d95a37a804-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367860 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d62103e-bb94-453d-b9cb-9a0de58823ce-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fwzk5\" (UID: \"3d62103e-bb94-453d-b9cb-9a0de58823ce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367901 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.367927 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-trusted-ca-bundle\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.368136 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-etcd-serving-ca\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.369006 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-trusted-ca\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.368997 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f37a4a33-7308-4546-a042-6b7335fbeccf-serving-cert\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.369269 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.369365 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f54afd7f-6b85-41b7-9a24-c253e16b8d88-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.369610 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/39830646-31e2-4717-9497-d499e89d16c3-auth-proxy-config\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.369811 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba7f02a-888d-4def-a53b-7a39407051fb-serving-cert\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.370031 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f40f7173-969d-45d1-baaa-8ed0eaab8499-serving-cert\") pod \"openshift-config-operator-7777fb866f-cj9n6\" (UID: \"f40f7173-969d-45d1-baaa-8ed0eaab8499\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.370225 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.370518 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.370702 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5cf156ab-e8de-4545-b432-dd538b3c7ce4-image-import-ca\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.371119 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.371365 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/209069f6-4788-4ce3-ac7f-6446f4c26e05-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.371818 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.371965 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.372731 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.373024 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/209069f6-4788-4ce3-ac7f-6446f4c26e05-etcd-client\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.373085 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5cf156ab-e8de-4545-b432-dd538b3c7ce4-etcd-client\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.373593 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d62103e-bb94-453d-b9cb-9a0de58823ce-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fwzk5\" (UID: \"3d62103e-bb94-453d-b9cb-9a0de58823ce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.374280 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4528d803-2051-4abb-9af2-898361f7b2fc-serving-cert\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.374371 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/209069f6-4788-4ce3-ac7f-6446f4c26e05-encryption-config\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.374506 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.374608 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/78c80454-cff3-4429-aafe-4359423f9b68-metrics-tls\") pod \"dns-operator-744455d44c-tw7jp\" (UID: \"78c80454-cff3-4429-aafe-4359423f9b68\") " pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.375002 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.376078 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n5bct"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.377179 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2374ca5a-3a6e-461b-9d9d-84d95a37a804-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.377663 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-p6m6d"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.377891 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.378335 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.378604 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.378893 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.379767 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.380430 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.382880 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n5bct"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.386937 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-p6m6d"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.388437 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-g5w9q"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.389055 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-g5w9q" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.390072 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-g5w9q"] Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.399920 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.421183 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.441574 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.460433 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.468878 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2353d56f-1cab-4181-b1c2-164a90efab5e-metrics-certs\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.469128 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kqt2\" (UniqueName: \"kubernetes.io/projected/fa028433-d9fc-4729-bdb0-65cd9a4695e5-kube-api-access-4kqt2\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.469322 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5wpt\" (UniqueName: \"kubernetes.io/projected/00cb67a6-0595-4968-a2d5-3a420bbd3c09-kube-api-access-m5wpt\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9h2w\" (UID: \"00cb67a6-0595-4968-a2d5-3a420bbd3c09\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.469430 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-oauth-serving-cert\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470054 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/805eb165-93ad-4085-92bc-28ad037ed23c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jx2z4\" (UID: \"805eb165-93ad-4085-92bc-28ad037ed23c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470112 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w75s\" (UniqueName: \"kubernetes.io/projected/13c92ba7-a34e-4f49-b98d-032d6d8543db-kube-api-access-5w75s\") pod \"downloads-7954f5f757-tvx5q\" (UID: \"13c92ba7-a34e-4f49-b98d-032d6d8543db\") " pod="openshift-console/downloads-7954f5f757-tvx5q" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470144 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/05d83c7d-74cd-4b0b-871e-4f522ed1775d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-htvkk\" (UID: \"05d83c7d-74cd-4b0b-871e-4f522ed1775d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470167 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2rstj\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470241 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05d83c7d-74cd-4b0b-871e-4f522ed1775d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-htvkk\" (UID: \"05d83c7d-74cd-4b0b-871e-4f522ed1775d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470275 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2353d56f-1cab-4181-b1c2-164a90efab5e-default-certificate\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470300 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-console-config\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470322 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m572w\" (UniqueName: \"kubernetes.io/projected/6467e871-dd98-4bcf-bf7c-53726589346b-kube-api-access-m572w\") pod \"marketplace-operator-79b997595-2rstj\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470356 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26b392c0-15c6-436c-9406-c410413c8f9d-serving-cert\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470381 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa028433-d9fc-4729-bdb0-65cd9a4695e5-proxy-tls\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470402 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/805eb165-93ad-4085-92bc-28ad037ed23c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jx2z4\" (UID: \"805eb165-93ad-4085-92bc-28ad037ed23c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470450 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-console-oauth-config\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470469 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2rstj\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470491 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/19cbbcfc-8187-4b59-a1e3-3a33e64e37c4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f6n6v\" (UID: \"19cbbcfc-8187-4b59-a1e3-3a33e64e37c4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470512 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/26b392c0-15c6-436c-9406-c410413c8f9d-etcd-service-ca\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470549 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgwc6\" (UniqueName: \"kubernetes.io/projected/26b392c0-15c6-436c-9406-c410413c8f9d-kube-api-access-rgwc6\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470567 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/26b392c0-15c6-436c-9406-c410413c8f9d-etcd-ca\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470591 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxr6d\" (UniqueName: \"kubernetes.io/projected/2353d56f-1cab-4181-b1c2-164a90efab5e-kube-api-access-xxr6d\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470621 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp9b4\" (UniqueName: \"kubernetes.io/projected/946f1f86-8ba8-4e6c-975a-b185265eb557-kube-api-access-rp9b4\") pod \"migrator-59844c95c7-krnvf\" (UID: \"946f1f86-8ba8-4e6c-975a-b185265eb557\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470641 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2353d56f-1cab-4181-b1c2-164a90efab5e-stats-auth\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470683 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/805eb165-93ad-4085-92bc-28ad037ed23c-config\") pod \"kube-apiserver-operator-766d6c64bb-jx2z4\" (UID: \"805eb165-93ad-4085-92bc-28ad037ed23c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470717 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-console-serving-cert\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470733 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cb67a6-0595-4968-a2d5-3a420bbd3c09-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9h2w\" (UID: \"00cb67a6-0595-4968-a2d5-3a420bbd3c09\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470753 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-service-ca\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470768 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1e0bb7e2-0204-4e78-afb5-6d0ecfd59541-srv-cert\") pod \"catalog-operator-68c6474976-5grcc\" (UID: \"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470792 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-trusted-ca-bundle\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470808 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fa028433-d9fc-4729-bdb0-65cd9a4695e5-images\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470831 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b392c0-15c6-436c-9406-c410413c8f9d-config\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470847 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/26b392c0-15c6-436c-9406-c410413c8f9d-etcd-client\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470862 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1e0bb7e2-0204-4e78-afb5-6d0ecfd59541-profile-collector-cert\") pod \"catalog-operator-68c6474976-5grcc\" (UID: \"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470880 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z4p9\" (UniqueName: \"kubernetes.io/projected/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-kube-api-access-7z4p9\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470903 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05d83c7d-74cd-4b0b-871e-4f522ed1775d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-htvkk\" (UID: \"05d83c7d-74cd-4b0b-871e-4f522ed1775d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470920 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4z6t\" (UniqueName: \"kubernetes.io/projected/19cbbcfc-8187-4b59-a1e3-3a33e64e37c4-kube-api-access-p4z6t\") pod \"multus-admission-controller-857f4d67dd-f6n6v\" (UID: \"19cbbcfc-8187-4b59-a1e3-3a33e64e37c4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470949 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2353d56f-1cab-4181-b1c2-164a90efab5e-service-ca-bundle\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.470971 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa028433-d9fc-4729-bdb0-65cd9a4695e5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.471068 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfbxx\" (UniqueName: \"kubernetes.io/projected/1e0bb7e2-0204-4e78-afb5-6d0ecfd59541-kube-api-access-rfbxx\") pod \"catalog-operator-68c6474976-5grcc\" (UID: \"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.471100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00cb67a6-0595-4968-a2d5-3a420bbd3c09-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9h2w\" (UID: \"00cb67a6-0595-4968-a2d5-3a420bbd3c09\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.471464 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-oauth-serving-cert\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.472371 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/26b392c0-15c6-436c-9406-c410413c8f9d-etcd-service-ca\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.472470 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-console-config\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.472518 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa028433-d9fc-4729-bdb0-65cd9a4695e5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.472983 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-service-ca\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.473432 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-trusted-ca-bundle\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.476416 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-console-oauth-config\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.476845 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-console-serving-cert\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.476850 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26b392c0-15c6-436c-9406-c410413c8f9d-serving-cert\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.479842 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.501171 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.511454 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7b27b5382a6bcb1d83f002e818fe625b57ec940e2422c4439b522487b5fcb4a3"} Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.511512 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/26b392c0-15c6-436c-9406-c410413c8f9d-etcd-ca\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.512874 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9b76dc3ad3a3e497ab2de4710e145a2d631837daf94f738f70f6cac20cf8683d"} Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.514690 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2655afb57de71a6aee9a1203954f43bc54bc23d661db6e7829e4001b431b5120"} Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.520214 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.522446 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26b392c0-15c6-436c-9406-c410413c8f9d-config\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.540108 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.561381 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.581055 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.587024 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-policies\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.600953 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.620334 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.639836 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.644552 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/26b392c0-15c6-436c-9406-c410413c8f9d-etcd-client\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.679944 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.699450 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.719953 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.740829 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.753575 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/805eb165-93ad-4085-92bc-28ad037ed23c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jx2z4\" (UID: \"805eb165-93ad-4085-92bc-28ad037ed23c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.760467 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.761895 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/805eb165-93ad-4085-92bc-28ad037ed23c-config\") pod \"kube-apiserver-operator-766d6c64bb-jx2z4\" (UID: \"805eb165-93ad-4085-92bc-28ad037ed23c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.780719 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.800050 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.806055 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05d83c7d-74cd-4b0b-871e-4f522ed1775d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-htvkk\" (UID: \"05d83c7d-74cd-4b0b-871e-4f522ed1775d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.820414 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.908703 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05d83c7d-74cd-4b0b-871e-4f522ed1775d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-htvkk\" (UID: \"05d83c7d-74cd-4b0b-871e-4f522ed1775d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.909590 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.909789 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.909928 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.910029 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.919275 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.940053 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.942312 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1e0bb7e2-0204-4e78-afb5-6d0ecfd59541-profile-collector-cert\") pod \"catalog-operator-68c6474976-5grcc\" (UID: \"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.944522 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/19cbbcfc-8187-4b59-a1e3-3a33e64e37c4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-f6n6v\" (UID: \"19cbbcfc-8187-4b59-a1e3-3a33e64e37c4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.960980 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.965138 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1e0bb7e2-0204-4e78-afb5-6d0ecfd59541-srv-cert\") pod \"catalog-operator-68c6474976-5grcc\" (UID: \"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:41 crc kubenswrapper[4715]: I1125 12:11:41.980314 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.001085 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.020933 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.024775 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2rstj\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.050331 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.051593 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2rstj\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.060380 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.080666 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.099644 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.120246 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.140414 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.159598 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.179672 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.200370 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.206137 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cb67a6-0595-4968-a2d5-3a420bbd3c09-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9h2w\" (UID: \"00cb67a6-0595-4968-a2d5-3a420bbd3c09\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.220026 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.222642 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00cb67a6-0595-4968-a2d5-3a420bbd3c09-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9h2w\" (UID: \"00cb67a6-0595-4968-a2d5-3a420bbd3c09\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.238075 4715 request.go:700] Waited for 1.009938335s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.241028 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.260382 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.280534 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.282319 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fa028433-d9fc-4729-bdb0-65cd9a4695e5-images\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.300097 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.304298 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2353d56f-1cab-4181-b1c2-164a90efab5e-default-certificate\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.320399 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.325102 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2353d56f-1cab-4181-b1c2-164a90efab5e-stats-auth\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.339878 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.360604 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.363961 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa028433-d9fc-4729-bdb0-65cd9a4695e5-proxy-tls\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.381049 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.399551 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.420160 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.433288 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2353d56f-1cab-4181-b1c2-164a90efab5e-metrics-certs\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.439540 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.442940 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2353d56f-1cab-4181-b1c2-164a90efab5e-service-ca-bundle\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.480445 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.500422 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.519595 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.519685 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3162e9cea8aa71c3eb3139387d8ab0ca75e6457aaeb5bb2b166549248dc260c3"} Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.521869 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ac1d12511ae60437e4c802dbdbd80edbf1039d2bc8d83df34802f5a7b803ca19"} Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.523249 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1323e36bb48dec38f80fefa083b9d4142e7410ab860a181f36c35baebfcd7e10"} Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.523387 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.539794 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.559983 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.580233 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.600154 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.625873 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.639919 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.659318 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.680730 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.700121 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.720894 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.740829 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.761089 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.780369 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.799638 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.820612 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.840062 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.859592 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.880383 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.900553 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.920509 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.956665 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn9tj\" (UniqueName: \"kubernetes.io/projected/78c80454-cff3-4429-aafe-4359423f9b68-kube-api-access-sn9tj\") pod \"dns-operator-744455d44c-tw7jp\" (UID: \"78c80454-cff3-4429-aafe-4359423f9b68\") " pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.975888 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-bound-sa-token\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:42 crc kubenswrapper[4715]: I1125 12:11:42.996399 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r2nn\" (UniqueName: \"kubernetes.io/projected/f37a4a33-7308-4546-a042-6b7335fbeccf-kube-api-access-2r2nn\") pod \"console-operator-58897d9998-5clnl\" (UID: \"f37a4a33-7308-4546-a042-6b7335fbeccf\") " pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.016084 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6kl5\" (UniqueName: \"kubernetes.io/projected/39830646-31e2-4717-9497-d499e89d16c3-kube-api-access-p6kl5\") pod \"machine-approver-56656f9798-4hg2w\" (UID: \"39830646-31e2-4717-9497-d499e89d16c3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.041473 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f54afd7f-6b85-41b7-9a24-c253e16b8d88-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.056885 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46btr\" (UniqueName: \"kubernetes.io/projected/209069f6-4788-4ce3-ac7f-6446f4c26e05-kube-api-access-46btr\") pod \"apiserver-7bbb656c7d-zmbnj\" (UID: \"209069f6-4788-4ce3-ac7f-6446f4c26e05\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.076764 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gznfl\" (UniqueName: \"kubernetes.io/projected/f606b97c-4f5c-4d2d-8f0d-b20fd861e719-kube-api-access-gznfl\") pod \"ingress-operator-5b745b69d9-956n7\" (UID: \"f606b97c-4f5c-4d2d-8f0d-b20fd861e719\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.078593 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" Nov 25 12:11:43 crc kubenswrapper[4715]: W1125 12:11:43.092252 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39830646_31e2_4717_9497_d499e89d16c3.slice/crio-5c13ea85e429b2af10fb3889ed8e03e47845eb567661427b3100fbb571a320c2 WatchSource:0}: Error finding container 5c13ea85e429b2af10fb3889ed8e03e47845eb567661427b3100fbb571a320c2: Status 404 returned error can't find the container with id 5c13ea85e429b2af10fb3889ed8e03e47845eb567661427b3100fbb571a320c2 Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.098764 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls47w\" (UniqueName: \"kubernetes.io/projected/bf92032d-fcbc-42cf-9740-ce8dbdde3a1e-kube-api-access-ls47w\") pod \"authentication-operator-69f744f599-7vcvm\" (UID: \"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.115706 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtdwp\" (UniqueName: \"kubernetes.io/projected/3d62103e-bb94-453d-b9cb-9a0de58823ce-kube-api-access-dtdwp\") pod \"openshift-controller-manager-operator-756b6f6bc6-fwzk5\" (UID: \"3d62103e-bb94-453d-b9cb-9a0de58823ce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.135852 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cmfq\" (UniqueName: \"kubernetes.io/projected/f40f7173-969d-45d1-baaa-8ed0eaab8499-kube-api-access-9cmfq\") pod \"openshift-config-operator-7777fb866f-cj9n6\" (UID: \"f40f7173-969d-45d1-baaa-8ed0eaab8499\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.136350 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.158468 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvtc6\" (UniqueName: \"kubernetes.io/projected/4528d803-2051-4abb-9af2-898361f7b2fc-kube-api-access-hvtc6\") pod \"controller-manager-879f6c89f-jtjqf\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.166592 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.176560 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n46p9\" (UniqueName: \"kubernetes.io/projected/b5c80b59-9196-4573-ac01-eb88a738fa25-kube-api-access-n46p9\") pod \"cluster-samples-operator-665b6dd947-qmkr2\" (UID: \"b5c80b59-9196-4573-ac01-eb88a738fa25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.185571 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.196024 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9fpp\" (UniqueName: \"kubernetes.io/projected/2374ca5a-3a6e-461b-9d9d-84d95a37a804-kube-api-access-d9fpp\") pod \"machine-api-operator-5694c8668f-4blqh\" (UID: \"2374ca5a-3a6e-461b-9d9d-84d95a37a804\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.211369 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.217112 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhkrj\" (UniqueName: \"kubernetes.io/projected/f54afd7f-6b85-41b7-9a24-c253e16b8d88-kube-api-access-xhkrj\") pod \"cluster-image-registry-operator-dc59b4c8b-zhxcr\" (UID: \"f54afd7f-6b85-41b7-9a24-c253e16b8d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.222431 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.238285 4715 request.go:700] Waited for 1.871249901s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.240033 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nttg6\" (UniqueName: \"kubernetes.io/projected/b2a3f84e-2d87-402d-8f01-c4965cd69dee-kube-api-access-nttg6\") pod \"openshift-apiserver-operator-796bbdcf4f-wmpt2\" (UID: \"b2a3f84e-2d87-402d-8f01-c4965cd69dee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.257353 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x45kf\" (UniqueName: \"kubernetes.io/projected/4ba7f02a-888d-4def-a53b-7a39407051fb-kube-api-access-x45kf\") pod \"route-controller-manager-6576b87f9c-m4kck\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.266120 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.277145 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmkz6\" (UniqueName: \"kubernetes.io/projected/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-kube-api-access-gmkz6\") pod \"oauth-openshift-558db77b4-qztrg\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.299952 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5f6v\" (UniqueName: \"kubernetes.io/projected/5cf156ab-e8de-4545-b432-dd538b3c7ce4-kube-api-access-n5f6v\") pod \"apiserver-76f77b778f-dl4rf\" (UID: \"5cf156ab-e8de-4545-b432-dd538b3c7ce4\") " pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.301144 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.321776 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.330460 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.341567 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.345266 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6"] Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.359650 4715 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.360679 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.362667 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.379755 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.386939 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5clnl"] Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.390522 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.401488 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 12:11:43 crc kubenswrapper[4715]: W1125 12:11:43.412241 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf37a4a33_7308_4546_a042_6b7335fbeccf.slice/crio-35e25a0c7dfdba0f79701efff52882867b6f6bcde60cc7411e61a73de7a5404a WatchSource:0}: Error finding container 35e25a0c7dfdba0f79701efff52882867b6f6bcde60cc7411e61a73de7a5404a: Status 404 returned error can't find the container with id 35e25a0c7dfdba0f79701efff52882867b6f6bcde60cc7411e61a73de7a5404a Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.414623 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.420111 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.440697 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.441423 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.452438 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tw7jp"] Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.459923 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.481902 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.494821 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.501862 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.523205 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kqt2\" (UniqueName: \"kubernetes.io/projected/fa028433-d9fc-4729-bdb0-65cd9a4695e5-kube-api-access-4kqt2\") pod \"machine-config-operator-74547568cd-bx96c\" (UID: \"fa028433-d9fc-4729-bdb0-65cd9a4695e5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.523392 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-956n7"] Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.533644 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" event={"ID":"39830646-31e2-4717-9497-d499e89d16c3","Type":"ContainerStarted","Data":"5b843881ae9ed44071ff774447dc768ac54dc74799f91f676a81c294f3f7b436"} Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.533695 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" event={"ID":"39830646-31e2-4717-9497-d499e89d16c3","Type":"ContainerStarted","Data":"5c13ea85e429b2af10fb3889ed8e03e47845eb567661427b3100fbb571a320c2"} Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.539098 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5wpt\" (UniqueName: \"kubernetes.io/projected/00cb67a6-0595-4968-a2d5-3a420bbd3c09-kube-api-access-m5wpt\") pod \"kube-storage-version-migrator-operator-b67b599dd-n9h2w\" (UID: \"00cb67a6-0595-4968-a2d5-3a420bbd3c09\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.542483 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" event={"ID":"f40f7173-969d-45d1-baaa-8ed0eaab8499","Type":"ContainerStarted","Data":"4d5f25b4578eac1070269e6dcf169f97a0592e464ba148a81cde9dcf41cad273"} Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.544832 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5clnl" event={"ID":"f37a4a33-7308-4546-a042-6b7335fbeccf","Type":"ContainerStarted","Data":"35e25a0c7dfdba0f79701efff52882867b6f6bcde60cc7411e61a73de7a5404a"} Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.546285 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" event={"ID":"78c80454-cff3-4429-aafe-4359423f9b68","Type":"ContainerStarted","Data":"a054693eabe9d91bd2961912a0aa72deb84318733552f41566d93c9788567c7a"} Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.558685 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/805eb165-93ad-4085-92bc-28ad037ed23c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jx2z4\" (UID: \"805eb165-93ad-4085-92bc-28ad037ed23c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.578478 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.582323 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/05d83c7d-74cd-4b0b-871e-4f522ed1775d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-htvkk\" (UID: \"05d83c7d-74cd-4b0b-871e-4f522ed1775d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.583894 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.588331 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.608987 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w75s\" (UniqueName: \"kubernetes.io/projected/13c92ba7-a34e-4f49-b98d-032d6d8543db-kube-api-access-5w75s\") pod \"downloads-7954f5f757-tvx5q\" (UID: \"13c92ba7-a34e-4f49-b98d-032d6d8543db\") " pod="openshift-console/downloads-7954f5f757-tvx5q" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.622502 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxr6d\" (UniqueName: \"kubernetes.io/projected/2353d56f-1cab-4181-b1c2-164a90efab5e-kube-api-access-xxr6d\") pod \"router-default-5444994796-vn6lh\" (UID: \"2353d56f-1cab-4181-b1c2-164a90efab5e\") " pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.633129 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.638585 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp9b4\" (UniqueName: \"kubernetes.io/projected/946f1f86-8ba8-4e6c-975a-b185265eb557-kube-api-access-rp9b4\") pod \"migrator-59844c95c7-krnvf\" (UID: \"946f1f86-8ba8-4e6c-975a-b185265eb557\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.645857 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.654050 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.659681 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.660156 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5"] Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.664767 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgwc6\" (UniqueName: \"kubernetes.io/projected/26b392c0-15c6-436c-9406-c410413c8f9d-kube-api-access-rgwc6\") pod \"etcd-operator-b45778765-zrx9b\" (UID: \"26b392c0-15c6-436c-9406-c410413c8f9d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.686254 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z4p9\" (UniqueName: \"kubernetes.io/projected/25b7b306-b4b9-4a46-bc4d-38df4f035c9f-kube-api-access-7z4p9\") pod \"console-f9d7485db-4dfdh\" (UID: \"25b7b306-b4b9-4a46-bc4d-38df4f035c9f\") " pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.705533 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4z6t\" (UniqueName: \"kubernetes.io/projected/19cbbcfc-8187-4b59-a1e3-3a33e64e37c4-kube-api-access-p4z6t\") pod \"multus-admission-controller-857f4d67dd-f6n6v\" (UID: \"19cbbcfc-8187-4b59-a1e3-3a33e64e37c4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.734092 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m572w\" (UniqueName: \"kubernetes.io/projected/6467e871-dd98-4bcf-bf7c-53726589346b-kube-api-access-m572w\") pod \"marketplace-operator-79b997595-2rstj\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.748902 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfbxx\" (UniqueName: \"kubernetes.io/projected/1e0bb7e2-0204-4e78-afb5-6d0ecfd59541-kube-api-access-rfbxx\") pod \"catalog-operator-68c6474976-5grcc\" (UID: \"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.768009 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7vcvm"] Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.811217 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj"] Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.830869 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tvx5q" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.839664 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.847707 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/deaac6b3-5004-4f5b-a7b1-26c51247fb9d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ttlfs\" (UID: \"deaac6b3-5004-4f5b-a7b1-26c51247fb9d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.847767 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d11b5ba8-bd0f-478e-92d3-e6de720040a4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-clmh7\" (UID: \"d11b5ba8-bd0f-478e-92d3-e6de720040a4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.847805 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d11b5ba8-bd0f-478e-92d3-e6de720040a4-srv-cert\") pod \"olm-operator-6b444d44fb-clmh7\" (UID: \"d11b5ba8-bd0f-478e-92d3-e6de720040a4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.847856 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/deaac6b3-5004-4f5b-a7b1-26c51247fb9d-proxy-tls\") pod \"machine-config-controller-84d6567774-ttlfs\" (UID: \"deaac6b3-5004-4f5b-a7b1-26c51247fb9d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.847889 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3088de81-03d3-483f-b6c1-ddf0a6e30c34-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.847918 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.847937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-tls\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.847968 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-995l6\" (UniqueName: \"kubernetes.io/projected/d11b5ba8-bd0f-478e-92d3-e6de720040a4-kube-api-access-995l6\") pod \"olm-operator-6b444d44fb-clmh7\" (UID: \"d11b5ba8-bd0f-478e-92d3-e6de720040a4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.847990 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nkvw\" (UniqueName: \"kubernetes.io/projected/deaac6b3-5004-4f5b-a7b1-26c51247fb9d-kube-api-access-7nkvw\") pod \"machine-config-controller-84d6567774-ttlfs\" (UID: \"deaac6b3-5004-4f5b-a7b1-26c51247fb9d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.848012 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-trusted-ca\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.848030 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-bound-sa-token\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.848048 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjrnk\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-kube-api-access-sjrnk\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.848080 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3088de81-03d3-483f-b6c1-ddf0a6e30c34-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.848107 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-certificates\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: E1125 12:11:43.849317 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.349299381 +0000 UTC m=+154.856802402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.850371 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.906155 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.918417 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.926645 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.950741 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.951361 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2"] Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.951448 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-995l6\" (UniqueName: \"kubernetes.io/projected/d11b5ba8-bd0f-478e-92d3-e6de720040a4-kube-api-access-995l6\") pod \"olm-operator-6b444d44fb-clmh7\" (UID: \"d11b5ba8-bd0f-478e-92d3-e6de720040a4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.951499 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nkvw\" (UniqueName: \"kubernetes.io/projected/deaac6b3-5004-4f5b-a7b1-26c51247fb9d-kube-api-access-7nkvw\") pod \"machine-config-controller-84d6567774-ttlfs\" (UID: \"deaac6b3-5004-4f5b-a7b1-26c51247fb9d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:43 crc kubenswrapper[4715]: E1125 12:11:43.951596 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.451562126 +0000 UTC m=+154.959065157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.951660 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5657bce1-46e6-4b26-b588-8de7a8c453ea-serving-cert\") pod \"service-ca-operator-777779d784-p6rsr\" (UID: \"5657bce1-46e6-4b26-b588-8de7a8c453ea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.951694 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dktf\" (UniqueName: \"kubernetes.io/projected/712c4b61-4bc2-41b0-8837-3d2d51e2a288-kube-api-access-7dktf\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.952356 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-trusted-ca\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.952398 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-bound-sa-token\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.952445 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjrnk\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-kube-api-access-sjrnk\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.952478 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/43155321-fb45-43dc-a204-07ab8b93d1a2-signing-key\") pod \"service-ca-9c57cc56f-wslh5\" (UID: \"43155321-fb45-43dc-a204-07ab8b93d1a2\") " pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.952927 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cce3ecd-396f-45bc-adf7-b323b6d427b7-config\") pod \"kube-controller-manager-operator-78b949d7b-t2dbv\" (UID: \"0cce3ecd-396f-45bc-adf7-b323b6d427b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.952975 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/829f6cba-1de5-4a2c-9f72-e24fe514caa9-config-volume\") pod \"collect-profiles-29401200-zr2x7\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.953096 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cce3ecd-396f-45bc-adf7-b323b6d427b7-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-t2dbv\" (UID: \"0cce3ecd-396f-45bc-adf7-b323b6d427b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.953252 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3088de81-03d3-483f-b6c1-ddf0a6e30c34-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.953349 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-trusted-ca\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.953369 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-certificates\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.953412 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/43155321-fb45-43dc-a204-07ab8b93d1a2-signing-cabundle\") pod \"service-ca-9c57cc56f-wslh5\" (UID: \"43155321-fb45-43dc-a204-07ab8b93d1a2\") " pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.953446 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-socket-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.953474 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/712c4b61-4bc2-41b0-8837-3d2d51e2a288-apiservice-cert\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.953527 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5657bce1-46e6-4b26-b588-8de7a8c453ea-config\") pod \"service-ca-operator-777779d784-p6rsr\" (UID: \"5657bce1-46e6-4b26-b588-8de7a8c453ea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.954108 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7d3445b9-20f1-4ab6-835a-80fea86c15b1-certs\") pod \"machine-config-server-2xgrz\" (UID: \"7d3445b9-20f1-4ab6-835a-80fea86c15b1\") " pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.954232 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7d3445b9-20f1-4ab6-835a-80fea86c15b1-node-bootstrap-token\") pod \"machine-config-server-2xgrz\" (UID: \"7d3445b9-20f1-4ab6-835a-80fea86c15b1\") " pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.954255 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghp55\" (UniqueName: \"kubernetes.io/projected/0c43525e-7149-48b7-aa59-94c4784e358b-kube-api-access-ghp55\") pod \"package-server-manager-789f6589d5-26dsn\" (UID: \"0c43525e-7149-48b7-aa59-94c4784e358b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.955158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/712c4b61-4bc2-41b0-8837-3d2d51e2a288-tmpfs\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.955242 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/deaac6b3-5004-4f5b-a7b1-26c51247fb9d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ttlfs\" (UID: \"deaac6b3-5004-4f5b-a7b1-26c51247fb9d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.955278 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d62ff1d-b062-4dfc-ba55-eae6dd68feaf-config-volume\") pod \"dns-default-p6m6d\" (UID: \"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf\") " pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.955413 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/45106950-65c8-40eb-8dd2-c51a9b82765b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8s95n\" (UID: \"45106950-65c8-40eb-8dd2-c51a9b82765b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.955478 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d11b5ba8-bd0f-478e-92d3-e6de720040a4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-clmh7\" (UID: \"d11b5ba8-bd0f-478e-92d3-e6de720040a4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.955587 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c43525e-7149-48b7-aa59-94c4784e358b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-26dsn\" (UID: \"0c43525e-7149-48b7-aa59-94c4784e358b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.955619 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3088de81-03d3-483f-b6c1-ddf0a6e30c34-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.955678 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d11b5ba8-bd0f-478e-92d3-e6de720040a4-srv-cert\") pod \"olm-operator-6b444d44fb-clmh7\" (UID: \"d11b5ba8-bd0f-478e-92d3-e6de720040a4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.955976 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvr7p\" (UniqueName: \"kubernetes.io/projected/7d3445b9-20f1-4ab6-835a-80fea86c15b1-kube-api-access-lvr7p\") pod \"machine-config-server-2xgrz\" (UID: \"7d3445b9-20f1-4ab6-835a-80fea86c15b1\") " pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.956056 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlwfk\" (UniqueName: \"kubernetes.io/projected/f5204538-5658-4148-8a88-7590e51114c5-kube-api-access-rlwfk\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.956305 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hhhx\" (UniqueName: \"kubernetes.io/projected/b75c7c74-8c79-4f30-af10-955c80ac8abb-kube-api-access-2hhhx\") pod \"ingress-canary-g5w9q\" (UID: \"b75c7c74-8c79-4f30-af10-955c80ac8abb\") " pod="openshift-ingress-canary/ingress-canary-g5w9q" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.956344 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-plugins-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.956381 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/deaac6b3-5004-4f5b-a7b1-26c51247fb9d-proxy-tls\") pod \"machine-config-controller-84d6567774-ttlfs\" (UID: \"deaac6b3-5004-4f5b-a7b1-26c51247fb9d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.956409 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptsz8\" (UniqueName: \"kubernetes.io/projected/4d62ff1d-b062-4dfc-ba55-eae6dd68feaf-kube-api-access-ptsz8\") pod \"dns-default-p6m6d\" (UID: \"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf\") " pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.956471 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-certificates\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.956488 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cce3ecd-396f-45bc-adf7-b323b6d427b7-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-t2dbv\" (UID: \"0cce3ecd-396f-45bc-adf7-b323b6d427b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.956532 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjkc5\" (UniqueName: \"kubernetes.io/projected/5657bce1-46e6-4b26-b588-8de7a8c453ea-kube-api-access-tjkc5\") pod \"service-ca-operator-777779d784-p6rsr\" (UID: \"5657bce1-46e6-4b26-b588-8de7a8c453ea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.959750 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j57r5\" (UniqueName: \"kubernetes.io/projected/43155321-fb45-43dc-a204-07ab8b93d1a2-kube-api-access-j57r5\") pod \"service-ca-9c57cc56f-wslh5\" (UID: \"43155321-fb45-43dc-a204-07ab8b93d1a2\") " pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.961603 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/829f6cba-1de5-4a2c-9f72-e24fe514caa9-secret-volume\") pod \"collect-profiles-29401200-zr2x7\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.961750 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/deaac6b3-5004-4f5b-a7b1-26c51247fb9d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ttlfs\" (UID: \"deaac6b3-5004-4f5b-a7b1-26c51247fb9d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.962780 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-mountpoint-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.963124 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3088de81-03d3-483f-b6c1-ddf0a6e30c34-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.963527 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4d62ff1d-b062-4dfc-ba55-eae6dd68feaf-metrics-tls\") pod \"dns-default-p6m6d\" (UID: \"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf\") " pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.963641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h64hb\" (UniqueName: \"kubernetes.io/projected/45106950-65c8-40eb-8dd2-c51a9b82765b-kube-api-access-h64hb\") pod \"control-plane-machine-set-operator-78cbb6b69f-8s95n\" (UID: \"45106950-65c8-40eb-8dd2-c51a9b82765b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.963740 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.963764 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-tls\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.964292 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d11b5ba8-bd0f-478e-92d3-e6de720040a4-srv-cert\") pod \"olm-operator-6b444d44fb-clmh7\" (UID: \"d11b5ba8-bd0f-478e-92d3-e6de720040a4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.964384 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75c7c74-8c79-4f30-af10-955c80ac8abb-cert\") pod \"ingress-canary-g5w9q\" (UID: \"b75c7c74-8c79-4f30-af10-955c80ac8abb\") " pod="openshift-ingress-canary/ingress-canary-g5w9q" Nov 25 12:11:43 crc kubenswrapper[4715]: E1125 12:11:43.964613 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.464598885 +0000 UTC m=+154.972101906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.964729 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-registration-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.964858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/712c4b61-4bc2-41b0-8837-3d2d51e2a288-webhook-cert\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.971564 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3088de81-03d3-483f-b6c1-ddf0a6e30c34-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.974266 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d11b5ba8-bd0f-478e-92d3-e6de720040a4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-clmh7\" (UID: \"d11b5ba8-bd0f-478e-92d3-e6de720040a4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.974904 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-tls\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.968412 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4b7q\" (UniqueName: \"kubernetes.io/projected/829f6cba-1de5-4a2c-9f72-e24fe514caa9-kube-api-access-b4b7q\") pod \"collect-profiles-29401200-zr2x7\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.978497 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-csi-data-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:43 crc kubenswrapper[4715]: I1125 12:11:43.983882 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/deaac6b3-5004-4f5b-a7b1-26c51247fb9d-proxy-tls\") pod \"machine-config-controller-84d6567774-ttlfs\" (UID: \"deaac6b3-5004-4f5b-a7b1-26c51247fb9d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.006552 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nkvw\" (UniqueName: \"kubernetes.io/projected/deaac6b3-5004-4f5b-a7b1-26c51247fb9d-kube-api-access-7nkvw\") pod \"machine-config-controller-84d6567774-ttlfs\" (UID: \"deaac6b3-5004-4f5b-a7b1-26c51247fb9d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.017388 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-995l6\" (UniqueName: \"kubernetes.io/projected/d11b5ba8-bd0f-478e-92d3-e6de720040a4-kube-api-access-995l6\") pod \"olm-operator-6b444d44fb-clmh7\" (UID: \"d11b5ba8-bd0f-478e-92d3-e6de720040a4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.025775 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4blqh"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.032396 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.038729 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-bound-sa-token\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.071983 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjrnk\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-kube-api-access-sjrnk\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.079383 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.579366664 +0000 UTC m=+155.086869685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079310 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079564 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4d62ff1d-b062-4dfc-ba55-eae6dd68feaf-metrics-tls\") pod \"dns-default-p6m6d\" (UID: \"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf\") " pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079595 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h64hb\" (UniqueName: \"kubernetes.io/projected/45106950-65c8-40eb-8dd2-c51a9b82765b-kube-api-access-h64hb\") pod \"control-plane-machine-set-operator-78cbb6b69f-8s95n\" (UID: \"45106950-65c8-40eb-8dd2-c51a9b82765b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079617 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079633 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75c7c74-8c79-4f30-af10-955c80ac8abb-cert\") pod \"ingress-canary-g5w9q\" (UID: \"b75c7c74-8c79-4f30-af10-955c80ac8abb\") " pod="openshift-ingress-canary/ingress-canary-g5w9q" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079652 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-registration-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079671 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/712c4b61-4bc2-41b0-8837-3d2d51e2a288-webhook-cert\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079690 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4b7q\" (UniqueName: \"kubernetes.io/projected/829f6cba-1de5-4a2c-9f72-e24fe514caa9-kube-api-access-b4b7q\") pod \"collect-profiles-29401200-zr2x7\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079704 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-csi-data-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079726 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5657bce1-46e6-4b26-b588-8de7a8c453ea-serving-cert\") pod \"service-ca-operator-777779d784-p6rsr\" (UID: \"5657bce1-46e6-4b26-b588-8de7a8c453ea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079756 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dktf\" (UniqueName: \"kubernetes.io/projected/712c4b61-4bc2-41b0-8837-3d2d51e2a288-kube-api-access-7dktf\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079773 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/43155321-fb45-43dc-a204-07ab8b93d1a2-signing-key\") pod \"service-ca-9c57cc56f-wslh5\" (UID: \"43155321-fb45-43dc-a204-07ab8b93d1a2\") " pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079792 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cce3ecd-396f-45bc-adf7-b323b6d427b7-config\") pod \"kube-controller-manager-operator-78b949d7b-t2dbv\" (UID: \"0cce3ecd-396f-45bc-adf7-b323b6d427b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079808 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/829f6cba-1de5-4a2c-9f72-e24fe514caa9-config-volume\") pod \"collect-profiles-29401200-zr2x7\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079826 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cce3ecd-396f-45bc-adf7-b323b6d427b7-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-t2dbv\" (UID: \"0cce3ecd-396f-45bc-adf7-b323b6d427b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079849 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/43155321-fb45-43dc-a204-07ab8b93d1a2-signing-cabundle\") pod \"service-ca-9c57cc56f-wslh5\" (UID: \"43155321-fb45-43dc-a204-07ab8b93d1a2\") " pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079871 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-socket-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079897 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/712c4b61-4bc2-41b0-8837-3d2d51e2a288-apiservice-cert\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079922 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5657bce1-46e6-4b26-b588-8de7a8c453ea-config\") pod \"service-ca-operator-777779d784-p6rsr\" (UID: \"5657bce1-46e6-4b26-b588-8de7a8c453ea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079946 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7d3445b9-20f1-4ab6-835a-80fea86c15b1-certs\") pod \"machine-config-server-2xgrz\" (UID: \"7d3445b9-20f1-4ab6-835a-80fea86c15b1\") " pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079964 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7d3445b9-20f1-4ab6-835a-80fea86c15b1-node-bootstrap-token\") pod \"machine-config-server-2xgrz\" (UID: \"7d3445b9-20f1-4ab6-835a-80fea86c15b1\") " pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079981 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghp55\" (UniqueName: \"kubernetes.io/projected/0c43525e-7149-48b7-aa59-94c4784e358b-kube-api-access-ghp55\") pod \"package-server-manager-789f6589d5-26dsn\" (UID: \"0c43525e-7149-48b7-aa59-94c4784e358b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.079996 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/712c4b61-4bc2-41b0-8837-3d2d51e2a288-tmpfs\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080014 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d62ff1d-b062-4dfc-ba55-eae6dd68feaf-config-volume\") pod \"dns-default-p6m6d\" (UID: \"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf\") " pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080033 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/45106950-65c8-40eb-8dd2-c51a9b82765b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8s95n\" (UID: \"45106950-65c8-40eb-8dd2-c51a9b82765b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080051 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c43525e-7149-48b7-aa59-94c4784e358b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-26dsn\" (UID: \"0c43525e-7149-48b7-aa59-94c4784e358b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080083 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvr7p\" (UniqueName: \"kubernetes.io/projected/7d3445b9-20f1-4ab6-835a-80fea86c15b1-kube-api-access-lvr7p\") pod \"machine-config-server-2xgrz\" (UID: \"7d3445b9-20f1-4ab6-835a-80fea86c15b1\") " pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080102 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlwfk\" (UniqueName: \"kubernetes.io/projected/f5204538-5658-4148-8a88-7590e51114c5-kube-api-access-rlwfk\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080123 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hhhx\" (UniqueName: \"kubernetes.io/projected/b75c7c74-8c79-4f30-af10-955c80ac8abb-kube-api-access-2hhhx\") pod \"ingress-canary-g5w9q\" (UID: \"b75c7c74-8c79-4f30-af10-955c80ac8abb\") " pod="openshift-ingress-canary/ingress-canary-g5w9q" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080143 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-plugins-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptsz8\" (UniqueName: \"kubernetes.io/projected/4d62ff1d-b062-4dfc-ba55-eae6dd68feaf-kube-api-access-ptsz8\") pod \"dns-default-p6m6d\" (UID: \"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf\") " pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080222 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cce3ecd-396f-45bc-adf7-b323b6d427b7-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-t2dbv\" (UID: \"0cce3ecd-396f-45bc-adf7-b323b6d427b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080241 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjkc5\" (UniqueName: \"kubernetes.io/projected/5657bce1-46e6-4b26-b588-8de7a8c453ea-kube-api-access-tjkc5\") pod \"service-ca-operator-777779d784-p6rsr\" (UID: \"5657bce1-46e6-4b26-b588-8de7a8c453ea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080259 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j57r5\" (UniqueName: \"kubernetes.io/projected/43155321-fb45-43dc-a204-07ab8b93d1a2-kube-api-access-j57r5\") pod \"service-ca-9c57cc56f-wslh5\" (UID: \"43155321-fb45-43dc-a204-07ab8b93d1a2\") " pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080279 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/829f6cba-1de5-4a2c-9f72-e24fe514caa9-secret-volume\") pod \"collect-profiles-29401200-zr2x7\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080304 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-mountpoint-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.080394 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-mountpoint-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.081785 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.581766463 +0000 UTC m=+155.089269554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.086266 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cce3ecd-396f-45bc-adf7-b323b6d427b7-config\") pod \"kube-controller-manager-operator-78b949d7b-t2dbv\" (UID: \"0cce3ecd-396f-45bc-adf7-b323b6d427b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.087174 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-registration-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.091167 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-csi-data-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.093375 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-socket-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.094103 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/829f6cba-1de5-4a2c-9f72-e24fe514caa9-config-volume\") pod \"collect-profiles-29401200-zr2x7\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.095025 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/712c4b61-4bc2-41b0-8837-3d2d51e2a288-tmpfs\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.095051 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f5204538-5658-4148-8a88-7590e51114c5-plugins-dir\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.095225 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5657bce1-46e6-4b26-b588-8de7a8c453ea-config\") pod \"service-ca-operator-777779d784-p6rsr\" (UID: \"5657bce1-46e6-4b26-b588-8de7a8c453ea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.096266 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4d62ff1d-b062-4dfc-ba55-eae6dd68feaf-metrics-tls\") pod \"dns-default-p6m6d\" (UID: \"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf\") " pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.096875 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d62ff1d-b062-4dfc-ba55-eae6dd68feaf-config-volume\") pod \"dns-default-p6m6d\" (UID: \"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf\") " pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.099030 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/712c4b61-4bc2-41b0-8837-3d2d51e2a288-webhook-cert\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.118953 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cce3ecd-396f-45bc-adf7-b323b6d427b7-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-t2dbv\" (UID: \"0cce3ecd-396f-45bc-adf7-b323b6d427b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.119631 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7d3445b9-20f1-4ab6-835a-80fea86c15b1-certs\") pod \"machine-config-server-2xgrz\" (UID: \"7d3445b9-20f1-4ab6-835a-80fea86c15b1\") " pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.120177 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/43155321-fb45-43dc-a204-07ab8b93d1a2-signing-key\") pod \"service-ca-9c57cc56f-wslh5\" (UID: \"43155321-fb45-43dc-a204-07ab8b93d1a2\") " pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.124050 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/712c4b61-4bc2-41b0-8837-3d2d51e2a288-apiservice-cert\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.124902 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c43525e-7149-48b7-aa59-94c4784e358b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-26dsn\" (UID: \"0c43525e-7149-48b7-aa59-94c4784e358b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.125476 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/43155321-fb45-43dc-a204-07ab8b93d1a2-signing-cabundle\") pod \"service-ca-9c57cc56f-wslh5\" (UID: \"43155321-fb45-43dc-a204-07ab8b93d1a2\") " pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.126065 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75c7c74-8c79-4f30-af10-955c80ac8abb-cert\") pod \"ingress-canary-g5w9q\" (UID: \"b75c7c74-8c79-4f30-af10-955c80ac8abb\") " pod="openshift-ingress-canary/ingress-canary-g5w9q" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.127509 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7d3445b9-20f1-4ab6-835a-80fea86c15b1-node-bootstrap-token\") pod \"machine-config-server-2xgrz\" (UID: \"7d3445b9-20f1-4ab6-835a-80fea86c15b1\") " pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.128746 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/45106950-65c8-40eb-8dd2-c51a9b82765b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-8s95n\" (UID: \"45106950-65c8-40eb-8dd2-c51a9b82765b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.137155 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5657bce1-46e6-4b26-b588-8de7a8c453ea-serving-cert\") pod \"service-ca-operator-777779d784-p6rsr\" (UID: \"5657bce1-46e6-4b26-b588-8de7a8c453ea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.138742 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/829f6cba-1de5-4a2c-9f72-e24fe514caa9-secret-volume\") pod \"collect-profiles-29401200-zr2x7\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.144943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4b7q\" (UniqueName: \"kubernetes.io/projected/829f6cba-1de5-4a2c-9f72-e24fe514caa9-kube-api-access-b4b7q\") pod \"collect-profiles-29401200-zr2x7\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.157172 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h64hb\" (UniqueName: \"kubernetes.io/projected/45106950-65c8-40eb-8dd2-c51a9b82765b-kube-api-access-h64hb\") pod \"control-plane-machine-set-operator-78cbb6b69f-8s95n\" (UID: \"45106950-65c8-40eb-8dd2-c51a9b82765b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.165833 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.167815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dktf\" (UniqueName: \"kubernetes.io/projected/712c4b61-4bc2-41b0-8837-3d2d51e2a288-kube-api-access-7dktf\") pod \"packageserver-d55dfcdfc-w4bkk\" (UID: \"712c4b61-4bc2-41b0-8837-3d2d51e2a288\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.182528 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.182761 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.682723251 +0000 UTC m=+155.190226272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.183900 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.184744 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.684732528 +0000 UTC m=+155.192235549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.188830 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cce3ecd-396f-45bc-adf7-b323b6d427b7-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-t2dbv\" (UID: \"0cce3ecd-396f-45bc-adf7-b323b6d427b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.194797 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.202919 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hhhx\" (UniqueName: \"kubernetes.io/projected/b75c7c74-8c79-4f30-af10-955c80ac8abb-kube-api-access-2hhhx\") pod \"ingress-canary-g5w9q\" (UID: \"b75c7c74-8c79-4f30-af10-955c80ac8abb\") " pod="openshift-ingress-canary/ingress-canary-g5w9q" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.224167 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jtjqf"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.226055 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghp55\" (UniqueName: \"kubernetes.io/projected/0c43525e-7149-48b7-aa59-94c4784e358b-kube-api-access-ghp55\") pod \"package-server-manager-789f6589d5-26dsn\" (UID: \"0c43525e-7149-48b7-aa59-94c4784e358b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.235762 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjkc5\" (UniqueName: \"kubernetes.io/projected/5657bce1-46e6-4b26-b588-8de7a8c453ea-kube-api-access-tjkc5\") pod \"service-ca-operator-777779d784-p6rsr\" (UID: \"5657bce1-46e6-4b26-b588-8de7a8c453ea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:44 crc kubenswrapper[4715]: W1125 12:11:44.252179 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4528d803_2051_4abb_9af2_898361f7b2fc.slice/crio-1aab59ec5e23cce40837389d62925f69919a475f2d3d00f3a38be31d8a85547a WatchSource:0}: Error finding container 1aab59ec5e23cce40837389d62925f69919a475f2d3d00f3a38be31d8a85547a: Status 404 returned error can't find the container with id 1aab59ec5e23cce40837389d62925f69919a475f2d3d00f3a38be31d8a85547a Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.262797 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvr7p\" (UniqueName: \"kubernetes.io/projected/7d3445b9-20f1-4ab6-835a-80fea86c15b1-kube-api-access-lvr7p\") pod \"machine-config-server-2xgrz\" (UID: \"7d3445b9-20f1-4ab6-835a-80fea86c15b1\") " pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.273607 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.274912 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.284682 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.287212 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.287710 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.787690772 +0000 UTC m=+155.295193793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.312120 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlwfk\" (UniqueName: \"kubernetes.io/projected/f5204538-5658-4148-8a88-7590e51114c5-kube-api-access-rlwfk\") pod \"csi-hostpathplugin-n5bct\" (UID: \"f5204538-5658-4148-8a88-7590e51114c5\") " pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.316860 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.324776 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptsz8\" (UniqueName: \"kubernetes.io/projected/4d62ff1d-b062-4dfc-ba55-eae6dd68feaf-kube-api-access-ptsz8\") pod \"dns-default-p6m6d\" (UID: \"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf\") " pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.331558 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.344927 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.359325 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j57r5\" (UniqueName: \"kubernetes.io/projected/43155321-fb45-43dc-a204-07ab8b93d1a2-kube-api-access-j57r5\") pod \"service-ca-9c57cc56f-wslh5\" (UID: \"43155321-fb45-43dc-a204-07ab8b93d1a2\") " pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.365343 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2xgrz" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.370825 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.389092 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.389520 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.889504526 +0000 UTC m=+155.397007547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.396083 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n5bct" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.401784 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qztrg"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.402029 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-g5w9q" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.404769 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.416255 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.422064 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.491895 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.492617 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:44.992597894 +0000 UTC m=+155.500100925 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.563607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-vn6lh" event={"ID":"2353d56f-1cab-4181-b1c2-164a90efab5e","Type":"ContainerStarted","Data":"05cacf956b3ce9df6782bc5795ecc18bee3c78f54936b9b34537d983c3878a8e"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.563659 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-vn6lh" event={"ID":"2353d56f-1cab-4181-b1c2-164a90efab5e","Type":"ContainerStarted","Data":"2c80ab5b45f5d7c6b671ada31d5bdcf34630f6548d3d2188b0c26556a4d1a07c"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.564731 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" event={"ID":"4528d803-2051-4abb-9af2-898361f7b2fc","Type":"ContainerStarted","Data":"1aab59ec5e23cce40837389d62925f69919a475f2d3d00f3a38be31d8a85547a"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.568136 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" event={"ID":"805eb165-93ad-4085-92bc-28ad037ed23c","Type":"ContainerStarted","Data":"19611d794fa6d5f081641ba4454de58c3e62a223d7f816429085c26b7e5707c3"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.572592 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" event={"ID":"f606b97c-4f5c-4d2d-8f0d-b20fd861e719","Type":"ContainerStarted","Data":"6681fef14dae2a0717cfc8e486c912e4a306e0b896adaec25e450b1ace94a746"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.572687 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" event={"ID":"f606b97c-4f5c-4d2d-8f0d-b20fd861e719","Type":"ContainerStarted","Data":"e177e48483c67ee66201adad1e72af2085ddea3c63059b2d397cf1d48ac0a997"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.596298 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.598351 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5clnl" event={"ID":"f37a4a33-7308-4546-a042-6b7335fbeccf","Type":"ContainerStarted","Data":"6eebdb73f2df2a011fef72f89f8498a1782a5d08122516ebfdd7e2e78ae9917f"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.599158 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.601847 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.101825026 +0000 UTC m=+155.609328047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.604307 4715 patch_prober.go:28] interesting pod/console-operator-58897d9998-5clnl container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.604359 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5clnl" podUID="f37a4a33-7308-4546-a042-6b7335fbeccf" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.605342 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.607871 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.607919 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" event={"ID":"b2a3f84e-2d87-402d-8f01-c4965cd69dee","Type":"ContainerStarted","Data":"98b9d977655c5b159481631762045b70cb11fcae2dd2da545314512c49694357"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.607946 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" event={"ID":"b2a3f84e-2d87-402d-8f01-c4965cd69dee","Type":"ContainerStarted","Data":"df260cc03491f4b61b475507e9cb6b2d1ff414a29b381232e91f7d95eb3fc0ba"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.613684 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dl4rf"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.613756 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" event={"ID":"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e","Type":"ContainerStarted","Data":"60f0fcc1100f6f75c6d22ab5a421c752aa6fb1d0af94db428653e67f176ec3f1"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.613785 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" event={"ID":"bf92032d-fcbc-42cf-9740-ce8dbdde3a1e","Type":"ContainerStarted","Data":"112de7e86ebfdb7f70de82cb8ab98f62d5d85fb7123e406ae752450309f6be96"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.617436 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.623711 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" event={"ID":"2374ca5a-3a6e-461b-9d9d-84d95a37a804","Type":"ContainerStarted","Data":"e55c15da8742a4592fc5a57b60ac18b257ea2ae0d38eb9a251f3e0325a31855f"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.623795 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" event={"ID":"2374ca5a-3a6e-461b-9d9d-84d95a37a804","Type":"ContainerStarted","Data":"61b0f2a7062c105d39f426dcc4f159636ed7a0f8743fb14fb89d32330874f5a6"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.631389 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" event={"ID":"3d62103e-bb94-453d-b9cb-9a0de58823ce","Type":"ContainerStarted","Data":"e0e952b8f75ee1664f035e9f625aa6820cb643f7c701507a303cf2714ea53ec8"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.631434 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" event={"ID":"3d62103e-bb94-453d-b9cb-9a0de58823ce","Type":"ContainerStarted","Data":"13cbbb7afb5a0a5b1f62b923676826df5120df894f255456640c656eeb8f406c"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.641746 4715 generic.go:334] "Generic (PLEG): container finished" podID="f40f7173-969d-45d1-baaa-8ed0eaab8499" containerID="ea718f9bc5b12282c1d52556be2142d8f207cc00b1248ac69f854743934cd7f9" exitCode=0 Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.641875 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" event={"ID":"f40f7173-969d-45d1-baaa-8ed0eaab8499","Type":"ContainerDied","Data":"ea718f9bc5b12282c1d52556be2142d8f207cc00b1248ac69f854743934cd7f9"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.646363 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" event={"ID":"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31","Type":"ContainerStarted","Data":"6049e232f320ad893bb680021fb2cf396b38d34972f10b6a9d44c80278debc0e"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.654742 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.658384 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk"] Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.660311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" event={"ID":"4ba7f02a-888d-4def-a53b-7a39407051fb","Type":"ContainerStarted","Data":"2bb9d9b41f11fd3398a587ab3f360da435c0e7950b560570c3715577655f3c56"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.661763 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.661810 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.663417 4715 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-m4kck container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.663542 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" podUID="4ba7f02a-888d-4def-a53b-7a39407051fb" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.665558 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" event={"ID":"39830646-31e2-4717-9497-d499e89d16c3","Type":"ContainerStarted","Data":"410699ac4794c72a6971acec14476b7885687d2e5cc0c82830a45ab511ffc10b"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.671570 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" event={"ID":"78c80454-cff3-4429-aafe-4359423f9b68","Type":"ContainerStarted","Data":"a47cb110fd715ccf7cc2e26b3d6dfb58a029fb1707467d94139cbd677ed0004d"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.683818 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" event={"ID":"209069f6-4788-4ce3-ac7f-6446f4c26e05","Type":"ContainerStarted","Data":"2e473c6a4417eaab71be2d07ef9461f22fca042dfebb9a568e48ee7b9441381d"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.683880 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" event={"ID":"209069f6-4788-4ce3-ac7f-6446f4c26e05","Type":"ContainerStarted","Data":"099dcbe6667fc20f211deb67981a681198eec57766fff5dd8528ac03cc8d6997"} Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.697211 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.700526 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.200496191 +0000 UTC m=+155.707999242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.753792 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.754133 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 25 12:11:44 crc kubenswrapper[4715]: W1125 12:11:44.776950 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf54afd7f_6b85_41b7_9a24_c253e16b8d88.slice/crio-5959dab1e9975a08209aa527fa24d3b5aebe48c371d6c45cf293dcb88e7cc7de WatchSource:0}: Error finding container 5959dab1e9975a08209aa527fa24d3b5aebe48c371d6c45cf293dcb88e7cc7de: Status 404 returned error can't find the container with id 5959dab1e9975a08209aa527fa24d3b5aebe48c371d6c45cf293dcb88e7cc7de Nov 25 12:11:44 crc kubenswrapper[4715]: W1125 12:11:44.787606 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05d83c7d_74cd_4b0b_871e_4f522ed1775d.slice/crio-31644246c4583afccfe0b70e167f66d7f74ecc8bdf4fa341543c93db8a9cc576 WatchSource:0}: Error finding container 31644246c4583afccfe0b70e167f66d7f74ecc8bdf4fa341543c93db8a9cc576: Status 404 returned error can't find the container with id 31644246c4583afccfe0b70e167f66d7f74ecc8bdf4fa341543c93db8a9cc576 Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.802470 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.804078 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.304060973 +0000 UTC m=+155.811564044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.876883 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-vn6lh" podStartSLOduration=127.876867614 podStartE2EDuration="2m7.876867614s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:44.876357909 +0000 UTC m=+155.383860950" watchObservedRunningTime="2025-11-25 12:11:44.876867614 +0000 UTC m=+155.384370635" Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.903403 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.903797 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.403743165 +0000 UTC m=+155.911246186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.903963 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:44 crc kubenswrapper[4715]: E1125 12:11:44.904306 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.40429164 +0000 UTC m=+155.911794661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:44 crc kubenswrapper[4715]: I1125 12:11:44.912980 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wmpt2" podStartSLOduration=128.912952895 podStartE2EDuration="2m8.912952895s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:44.910887817 +0000 UTC m=+155.418390828" watchObservedRunningTime="2025-11-25 12:11:44.912952895 +0000 UTC m=+155.420455916" Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.005084 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.005843 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.505814665 +0000 UTC m=+156.013317686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.108912 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.109502 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.60948258 +0000 UTC m=+156.116985601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.120897 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc"] Nov 25 12:11:45 crc kubenswrapper[4715]: W1125 12:11:45.193488 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e0bb7e2_0204_4e78_afb5_6d0ecfd59541.slice/crio-3d3bbe6403632ac01e591878856669cc9cd92c928b6471db5694205d619aa394 WatchSource:0}: Error finding container 3d3bbe6403632ac01e591878856669cc9cd92c928b6471db5694205d619aa394: Status 404 returned error can't find the container with id 3d3bbe6403632ac01e591878856669cc9cd92c928b6471db5694205d619aa394 Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.212705 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.213210 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.713176796 +0000 UTC m=+156.220679817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.213368 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.213710 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.713700361 +0000 UTC m=+156.221203382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.305863 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tvx5q"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.314352 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.314788 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.814775103 +0000 UTC m=+156.322278124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.326557 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4dfdh"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.347063 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hg2w" podStartSLOduration=129.347045516 podStartE2EDuration="2m9.347045516s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:45.344649549 +0000 UTC m=+155.852152580" watchObservedRunningTime="2025-11-25 12:11:45.347045516 +0000 UTC m=+155.854548537" Nov 25 12:11:45 crc kubenswrapper[4715]: W1125 12:11:45.367377 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25b7b306_b4b9_4a46_bc4d_38df4f035c9f.slice/crio-0e41a6d3163e26df143b6bfb5af9e6e617a277bbe3332bf82b4c355dbc20687e WatchSource:0}: Error finding container 0e41a6d3163e26df143b6bfb5af9e6e617a277bbe3332bf82b4c355dbc20687e: Status 404 returned error can't find the container with id 0e41a6d3163e26df143b6bfb5af9e6e617a277bbe3332bf82b4c355dbc20687e Nov 25 12:11:45 crc kubenswrapper[4715]: W1125 12:11:45.390198 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13c92ba7_a34e_4f49_b98d_032d6d8543db.slice/crio-bf2a80f2cf7b7cdc7523b95b6f5eadb4c39030da25fd767e378843f88fd4d43a WatchSource:0}: Error finding container bf2a80f2cf7b7cdc7523b95b6f5eadb4c39030da25fd767e378843f88fd4d43a: Status 404 returned error can't find the container with id bf2a80f2cf7b7cdc7523b95b6f5eadb4c39030da25fd767e378843f88fd4d43a Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.416539 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zrx9b"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.435875 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" podStartSLOduration=128.43584761 podStartE2EDuration="2m8.43584761s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:45.400124399 +0000 UTC m=+155.907627430" watchObservedRunningTime="2025-11-25 12:11:45.43584761 +0000 UTC m=+155.943350631" Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.437231 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.438046 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:45.938032002 +0000 UTC m=+156.445535023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.461582 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-f6n6v"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.475376 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.488508 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-7vcvm" podStartSLOduration=129.488485441 podStartE2EDuration="2m9.488485441s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:45.476256175 +0000 UTC m=+155.983759216" watchObservedRunningTime="2025-11-25 12:11:45.488485441 +0000 UTC m=+155.995988462" Nov 25 12:11:45 crc kubenswrapper[4715]: W1125 12:11:45.508024 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26b392c0_15c6_436c_9406_c410413c8f9d.slice/crio-4f2ff809e9925b9a26d2179986b097ad31e31e95caaac5f5b3d3890d171e4298 WatchSource:0}: Error finding container 4f2ff809e9925b9a26d2179986b097ad31e31e95caaac5f5b3d3890d171e4298: Status 404 returned error can't find the container with id 4f2ff809e9925b9a26d2179986b097ad31e31e95caaac5f5b3d3890d171e4298 Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.541901 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.545470 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7"] Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.545478 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.045452933 +0000 UTC m=+156.552956014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.554425 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.600431 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-5clnl" podStartSLOduration=129.60041015 podStartE2EDuration="2m9.60041015s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:45.598750032 +0000 UTC m=+156.106253063" watchObservedRunningTime="2025-11-25 12:11:45.60041015 +0000 UTC m=+156.107913171" Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.610860 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.631346 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.646498 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.646912 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.146896125 +0000 UTC m=+156.654399146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.647704 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rstj"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.665790 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.676910 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:45 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:45 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:45 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.677002 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.690958 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.696957 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-p6m6d"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.716906 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wslh5"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.742623 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" event={"ID":"4ba7f02a-888d-4def-a53b-7a39407051fb","Type":"ContainerStarted","Data":"fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.744721 4715 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-m4kck container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.744777 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" podUID="4ba7f02a-888d-4def-a53b-7a39407051fb" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.745210 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n5bct"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.748024 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-g5w9q"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.751381 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.751957 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.25193764 +0000 UTC m=+156.759440661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.752029 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.755741 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.255720537 +0000 UTC m=+156.763223558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.759526 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn"] Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.768108 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" event={"ID":"2374ca5a-3a6e-461b-9d9d-84d95a37a804","Type":"ContainerStarted","Data":"3d97add1e3cdfb88a769eb147620d0401b5d9af003c79b839f2d959810d9d3a7"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.785211 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" event={"ID":"deaac6b3-5004-4f5b-a7b1-26c51247fb9d","Type":"ContainerStarted","Data":"770b642c113e45c0d6ac50e7ff4669c178a0aac8c70bf83e361ef713bdd5d6bf"} Nov 25 12:11:45 crc kubenswrapper[4715]: W1125 12:11:45.786986 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d62ff1d_b062_4dfc_ba55_eae6dd68feaf.slice/crio-2ed53e8350f973a51d2ca80131b1cd7c0cd089671407961fbd34c9094be4982b WatchSource:0}: Error finding container 2ed53e8350f973a51d2ca80131b1cd7c0cd089671407961fbd34c9094be4982b: Status 404 returned error can't find the container with id 2ed53e8350f973a51d2ca80131b1cd7c0cd089671407961fbd34c9094be4982b Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.790764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" event={"ID":"45106950-65c8-40eb-8dd2-c51a9b82765b","Type":"ContainerStarted","Data":"10f9b0681555eeac633a59a8c66c7447ba90f690a64606b751546f6be24fd77b"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.798539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" event={"ID":"26b392c0-15c6-436c-9406-c410413c8f9d","Type":"ContainerStarted","Data":"4f2ff809e9925b9a26d2179986b097ad31e31e95caaac5f5b3d3890d171e4298"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.800077 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" event={"ID":"fa028433-d9fc-4729-bdb0-65cd9a4695e5","Type":"ContainerStarted","Data":"86615193c3177c29007c8f44573297b0c833401107a64421e574f380e950e9a7"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.806989 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" event={"ID":"b5c80b59-9196-4573-ac01-eb88a738fa25","Type":"ContainerStarted","Data":"e031532b83a208e5ce2df207faa53f635f44da28c67efe9c6529e5264ebbb165"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.810213 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" event={"ID":"5cf156ab-e8de-4545-b432-dd538b3c7ce4","Type":"ContainerStarted","Data":"1166fc9a241da506a138736e64663c9c971a1de6ca02ceaeff275883ba764e18"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.812077 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" event={"ID":"f54afd7f-6b85-41b7-9a24-c253e16b8d88","Type":"ContainerStarted","Data":"5959dab1e9975a08209aa527fa24d3b5aebe48c371d6c45cf293dcb88e7cc7de"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.815340 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" event={"ID":"f606b97c-4f5c-4d2d-8f0d-b20fd861e719","Type":"ContainerStarted","Data":"707d496282ba0072b5923cc945c30b1f3c95863dd0cc5795b5c550dbf30504b7"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.818124 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tvx5q" event={"ID":"13c92ba7-a34e-4f49-b98d-032d6d8543db","Type":"ContainerStarted","Data":"bf2a80f2cf7b7cdc7523b95b6f5eadb4c39030da25fd767e378843f88fd4d43a"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.821562 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" event={"ID":"f40f7173-969d-45d1-baaa-8ed0eaab8499","Type":"ContainerStarted","Data":"d3b0bc3881cd451eff05a3fc9f0904f79c3ce98f1ddf642002f80ed7b1eb8744"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.822472 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.833866 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" event={"ID":"00cb67a6-0595-4968-a2d5-3a420bbd3c09","Type":"ContainerStarted","Data":"cb318ba6a169a52efdc96801d80f85d46f9071be3ab05c98761cb4d89470f8e6"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.833911 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" event={"ID":"00cb67a6-0595-4968-a2d5-3a420bbd3c09","Type":"ContainerStarted","Data":"62696d0ef2a7635ca9e0e658a33a751fc65992cb64b5b80678f677ff01a07af9"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.845467 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" event={"ID":"05d83c7d-74cd-4b0b-871e-4f522ed1775d","Type":"ContainerStarted","Data":"31644246c4583afccfe0b70e167f66d7f74ecc8bdf4fa341543c93db8a9cc576"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.853253 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.855096 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.35507867 +0000 UTC m=+156.862581691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.857353 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" event={"ID":"d11b5ba8-bd0f-478e-92d3-e6de720040a4","Type":"ContainerStarted","Data":"970d6ca908df6eb24644046c421b95741ea6c12976f9765a4a2f3fc9deca1658"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.860524 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4dfdh" event={"ID":"25b7b306-b4b9-4a46-bc4d-38df4f035c9f","Type":"ContainerStarted","Data":"0e41a6d3163e26df143b6bfb5af9e6e617a277bbe3332bf82b4c355dbc20687e"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.863168 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" event={"ID":"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541","Type":"ContainerStarted","Data":"3d3bbe6403632ac01e591878856669cc9cd92c928b6471db5694205d619aa394"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.868758 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2xgrz" event={"ID":"7d3445b9-20f1-4ab6-835a-80fea86c15b1","Type":"ContainerStarted","Data":"f341555f3e75f143f654e1ee9b98e12f17668e466b4006a2add181f994d61873"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.868811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2xgrz" event={"ID":"7d3445b9-20f1-4ab6-835a-80fea86c15b1","Type":"ContainerStarted","Data":"6e6da1288129cfda6a4188c6001011f559fb42362f9375b7b8b709d5ddf552e8"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.874060 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" event={"ID":"0cce3ecd-396f-45bc-adf7-b323b6d427b7","Type":"ContainerStarted","Data":"0489f378929fd2c0ba6d7a1824988ae72a1f7bde14cb26fcad4242fe5ca8de46"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.876110 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" event={"ID":"78c80454-cff3-4429-aafe-4359423f9b68","Type":"ContainerStarted","Data":"6f3ccc51e3777e28616405730401a806322293d32ba3d008ec8c2ae763be92aa"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.880463 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" event={"ID":"829f6cba-1de5-4a2c-9f72-e24fe514caa9","Type":"ContainerStarted","Data":"cd2a32af04e2aeb42bad1a446ec3fe2309e1bee6e9162ed05ec767b91c092131"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.918443 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" event={"ID":"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31","Type":"ContainerStarted","Data":"822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.919507 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.933932 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" event={"ID":"19cbbcfc-8187-4b59-a1e3-3a33e64e37c4","Type":"ContainerStarted","Data":"c499cdeb1121d5cb24ec41b2d4e992e295e0f7d98c6c2d4139239a7d1bafb690"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.939689 4715 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-qztrg container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" start-of-body= Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.939768 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" podUID="09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.955665 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" event={"ID":"805eb165-93ad-4085-92bc-28ad037ed23c","Type":"ContainerStarted","Data":"6a39b93536da987ea50c5feaa70d8466d5f62acf3d52a670022cdc41a2981257"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.958052 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:45 crc kubenswrapper[4715]: E1125 12:11:45.961240 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.461218706 +0000 UTC m=+156.968721737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.965149 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf" event={"ID":"946f1f86-8ba8-4e6c-975a-b185265eb557","Type":"ContainerStarted","Data":"171016bb28a8eb0875119c24193e51ed24099b6f2f02046209f3841fa6abf87e"} Nov 25 12:11:45 crc kubenswrapper[4715]: I1125 12:11:45.965238 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf" event={"ID":"946f1f86-8ba8-4e6c-975a-b185265eb557","Type":"ContainerStarted","Data":"b10e2163798d83a8ed1dd7e0f6aabddb426b4189da8a808e68c53c3b40637535"} Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.000061 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" event={"ID":"4528d803-2051-4abb-9af2-898361f7b2fc","Type":"ContainerStarted","Data":"68050fbf9456c16612dc5eb6d2607e0774dedd8e25cfe7bb2708bfb88d332df2"} Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.001092 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.017427 4715 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-jtjqf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.017483 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" podUID="4528d803-2051-4abb-9af2-898361f7b2fc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.024759 4715 generic.go:334] "Generic (PLEG): container finished" podID="209069f6-4788-4ce3-ac7f-6446f4c26e05" containerID="2e473c6a4417eaab71be2d07ef9461f22fca042dfebb9a568e48ee7b9441381d" exitCode=0 Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.026252 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" event={"ID":"209069f6-4788-4ce3-ac7f-6446f4c26e05","Type":"ContainerDied","Data":"2e473c6a4417eaab71be2d07ef9461f22fca042dfebb9a568e48ee7b9441381d"} Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.026501 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" event={"ID":"209069f6-4788-4ce3-ac7f-6446f4c26e05","Type":"ContainerStarted","Data":"92204ca4030a4db13875b4c23221acbfac9332a774fc46daea176cf126200ce8"} Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.045941 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-5clnl" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.059810 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.061125 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.561104743 +0000 UTC m=+157.068607764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.156499 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" podStartSLOduration=130.156479813 podStartE2EDuration="2m10.156479813s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.155142916 +0000 UTC m=+156.662645937" watchObservedRunningTime="2025-11-25 12:11:46.156479813 +0000 UTC m=+156.663982834" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.162773 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fwzk5" podStartSLOduration=130.162746381 podStartE2EDuration="2m10.162746381s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.127490983 +0000 UTC m=+156.634994004" watchObservedRunningTime="2025-11-25 12:11:46.162746381 +0000 UTC m=+156.670249412" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.163018 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.169289 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.669273826 +0000 UTC m=+157.176776847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.205224 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-956n7" podStartSLOduration=129.205202343 podStartE2EDuration="2m9.205202343s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.20191496 +0000 UTC m=+156.709417981" watchObservedRunningTime="2025-11-25 12:11:46.205202343 +0000 UTC m=+156.712705364" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.255768 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2xgrz" podStartSLOduration=5.2557533339999996 podStartE2EDuration="5.255753334s" podCreationTimestamp="2025-11-25 12:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.254930791 +0000 UTC m=+156.762433812" watchObservedRunningTime="2025-11-25 12:11:46.255753334 +0000 UTC m=+156.763256355" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.271076 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.271528 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.77150302 +0000 UTC m=+157.279006081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.287603 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-tw7jp" podStartSLOduration=129.287584685 podStartE2EDuration="2m9.287584685s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.283375567 +0000 UTC m=+156.790878578" watchObservedRunningTime="2025-11-25 12:11:46.287584685 +0000 UTC m=+156.795087706" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.377165 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.377520 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.877508261 +0000 UTC m=+157.385011282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.399644 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" podStartSLOduration=130.399623708 podStartE2EDuration="2m10.399623708s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.368781824 +0000 UTC m=+156.876284845" watchObservedRunningTime="2025-11-25 12:11:46.399623708 +0000 UTC m=+156.907126739" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.431675 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" podStartSLOduration=130.431656395 podStartE2EDuration="2m10.431656395s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.402159689 +0000 UTC m=+156.909662710" watchObservedRunningTime="2025-11-25 12:11:46.431656395 +0000 UTC m=+156.939159416" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.465869 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" podStartSLOduration=130.465847743 podStartE2EDuration="2m10.465847743s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.433877747 +0000 UTC m=+156.941380768" watchObservedRunningTime="2025-11-25 12:11:46.465847743 +0000 UTC m=+156.973350764" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.480238 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.482223 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.982177375 +0000 UTC m=+157.489680466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.483430 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.485827 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:46.985810878 +0000 UTC m=+157.493313899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.498445 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-4blqh" podStartSLOduration=129.498425375 podStartE2EDuration="2m9.498425375s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.468361004 +0000 UTC m=+156.975864025" watchObservedRunningTime="2025-11-25 12:11:46.498425375 +0000 UTC m=+157.005928406" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.542699 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" podStartSLOduration=129.542680388 podStartE2EDuration="2m9.542680388s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.541753352 +0000 UTC m=+157.049256383" watchObservedRunningTime="2025-11-25 12:11:46.542680388 +0000 UTC m=+157.050183419" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.571602 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jx2z4" podStartSLOduration=129.571581227 podStartE2EDuration="2m9.571581227s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.570766493 +0000 UTC m=+157.078269514" watchObservedRunningTime="2025-11-25 12:11:46.571581227 +0000 UTC m=+157.079084248" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.589009 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.589503 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:47.089454252 +0000 UTC m=+157.596957273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.664842 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:46 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:46 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:46 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.665117 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.691089 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.691458 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:47.19144647 +0000 UTC m=+157.698949491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.794069 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.794461 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:47.294442926 +0000 UTC m=+157.801945947 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.896355 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.896801 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:47.396786664 +0000 UTC m=+157.904289695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:46 crc kubenswrapper[4715]: I1125 12:11:46.997645 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:46 crc kubenswrapper[4715]: E1125 12:11:46.998616 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:47.498596556 +0000 UTC m=+158.006099587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.100066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.100457 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:47.600442169 +0000 UTC m=+158.107945190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.104633 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" event={"ID":"d11b5ba8-bd0f-478e-92d3-e6de720040a4","Type":"ContainerStarted","Data":"40d91556fedad24635b024d76353e3f28b0140b59b5637852594a5c8c1ab2012"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.105570 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.106773 4715 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-clmh7 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.106826 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" podUID="d11b5ba8-bd0f-478e-92d3-e6de720040a4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.127490 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" event={"ID":"05d83c7d-74cd-4b0b-871e-4f522ed1775d","Type":"ContainerStarted","Data":"51f9c0007a8fc53e7be2ddc41bb91ce9a797b0ae1ee7266327916d68529ce67e"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.143370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-p6m6d" event={"ID":"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf","Type":"ContainerStarted","Data":"2ed53e8350f973a51d2ca80131b1cd7c0cd089671407961fbd34c9094be4982b"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.181045 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" podStartSLOduration=130.181025561 podStartE2EDuration="2m10.181025561s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.178275573 +0000 UTC m=+157.685778604" watchObservedRunningTime="2025-11-25 12:11:47.181025561 +0000 UTC m=+157.688528582" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.182682 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" event={"ID":"0cce3ecd-396f-45bc-adf7-b323b6d427b7","Type":"ContainerStarted","Data":"d9722c1d92b8381de8c30278648f02895063567c6c70ccf758d1a999817e18f0"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.184275 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n9h2w" podStartSLOduration=130.184254233 podStartE2EDuration="2m10.184254233s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:46.604400635 +0000 UTC m=+157.111903666" watchObservedRunningTime="2025-11-25 12:11:47.184254233 +0000 UTC m=+157.691757254" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.202002 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.204547 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-htvkk" podStartSLOduration=130.204531076 podStartE2EDuration="2m10.204531076s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.201627274 +0000 UTC m=+157.709130295" watchObservedRunningTime="2025-11-25 12:11:47.204531076 +0000 UTC m=+157.712034097" Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.205709 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:47.70568563 +0000 UTC m=+158.213188701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.235048 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" event={"ID":"6467e871-dd98-4bcf-bf7c-53726589346b","Type":"ContainerStarted","Data":"c9a264aae6a04008ac0b82b97a804163fb63cd6646481d94e62aeaeb9d9a9ce9"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.242480 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-t2dbv" podStartSLOduration=130.242463311 podStartE2EDuration="2m10.242463311s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.240735192 +0000 UTC m=+157.748238213" watchObservedRunningTime="2025-11-25 12:11:47.242463311 +0000 UTC m=+157.749966332" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.252375 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhxcr" event={"ID":"f54afd7f-6b85-41b7-9a24-c253e16b8d88","Type":"ContainerStarted","Data":"00d2e52c64f9873b0a4a4f91e79ef77eb52dde38ee357d1503c248397c8c5397"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.262106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" event={"ID":"712c4b61-4bc2-41b0-8837-3d2d51e2a288","Type":"ContainerStarted","Data":"8f1752fcaa9a1d81cedfa031b665b701132cd8bf6030f99d9bdf53c6766f2ba2"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.262460 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" event={"ID":"712c4b61-4bc2-41b0-8837-3d2d51e2a288","Type":"ContainerStarted","Data":"585ac51698c7ef936950fe92bb7e56de72c08631ee2e8fa6f48124f89eb65d10"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.263588 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.269054 4715 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-w4bkk container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.269113 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" podUID="712c4b61-4bc2-41b0-8837-3d2d51e2a288" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.273355 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" event={"ID":"19cbbcfc-8187-4b59-a1e3-3a33e64e37c4","Type":"ContainerStarted","Data":"c1776ec72764538dc4855f0d3895e8169d4bf0429edc90f269c3be7542070dbc"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.283655 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" event={"ID":"fa028433-d9fc-4729-bdb0-65cd9a4695e5","Type":"ContainerStarted","Data":"da4e5533fe1e4500f7c3469c8698aff4eedb24d1d636fc82412ee7dd62f98f9f"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.297881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5bct" event={"ID":"f5204538-5658-4148-8a88-7590e51114c5","Type":"ContainerStarted","Data":"15fa11c85da365db37a3af4017f17ac98a7c74ea4f94a775c12e053f7d05b2d3"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.301714 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" podStartSLOduration=130.301698968 podStartE2EDuration="2m10.301698968s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.300170805 +0000 UTC m=+157.807673826" watchObservedRunningTime="2025-11-25 12:11:47.301698968 +0000 UTC m=+157.809201989" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.303990 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.305605 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:47.805593579 +0000 UTC m=+158.313096600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.311734 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" event={"ID":"829f6cba-1de5-4a2c-9f72-e24fe514caa9","Type":"ContainerStarted","Data":"4210d8fd6ab3f8692a875f90e733091d229bb410a2edc91256c094a426cf95f0"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.329844 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" podStartSLOduration=130.329827524 podStartE2EDuration="2m10.329827524s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.329200587 +0000 UTC m=+157.836703608" watchObservedRunningTime="2025-11-25 12:11:47.329827524 +0000 UTC m=+157.837330545" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.330210 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" event={"ID":"b5c80b59-9196-4573-ac01-eb88a738fa25","Type":"ContainerStarted","Data":"b6af6f74c55d2db4a9e428979f6e0aef45faf9e6d2c01164efed02e6555851b2"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.350983 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" event={"ID":"43155321-fb45-43dc-a204-07ab8b93d1a2","Type":"ContainerStarted","Data":"6c099038bcce830b8b8583983954d6bba879b3c2f61664274e4d5bf58a50512d"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.351054 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" event={"ID":"43155321-fb45-43dc-a204-07ab8b93d1a2","Type":"ContainerStarted","Data":"6e1a4639628701bf0781eb8161c1e3e5e933d9bbedd521dc02ef776bfa2aa4ca"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.357648 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" podStartSLOduration=131.357635232 podStartE2EDuration="2m11.357635232s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.35720933 +0000 UTC m=+157.864712351" watchObservedRunningTime="2025-11-25 12:11:47.357635232 +0000 UTC m=+157.865138253" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.374616 4715 generic.go:334] "Generic (PLEG): container finished" podID="5cf156ab-e8de-4545-b432-dd538b3c7ce4" containerID="f9f568e5fa66cd259a2846b6d512e75c6653ca1a8684609895f1f7cf715d6055" exitCode=0 Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.374729 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" event={"ID":"5cf156ab-e8de-4545-b432-dd538b3c7ce4","Type":"ContainerDied","Data":"f9f568e5fa66cd259a2846b6d512e75c6653ca1a8684609895f1f7cf715d6055"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.390914 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" event={"ID":"5657bce1-46e6-4b26-b588-8de7a8c453ea","Type":"ContainerStarted","Data":"5e3537e4dd1e4425c16eaecbcbcd7a1ddf62a451338c985d9073ef53ef67141e"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.390961 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" event={"ID":"5657bce1-46e6-4b26-b588-8de7a8c453ea","Type":"ContainerStarted","Data":"097e96147d53e14ad093f5104f7d162a0eb565f6dd527bc7cde9c7f9cfc6aaaf"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.392690 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" podStartSLOduration=131.392672754 podStartE2EDuration="2m11.392672754s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.39183426 +0000 UTC m=+157.899337281" watchObservedRunningTime="2025-11-25 12:11:47.392672754 +0000 UTC m=+157.900175785" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.405730 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" event={"ID":"26b392c0-15c6-436c-9406-c410413c8f9d","Type":"ContainerStarted","Data":"45d2d1a5859bcd5b83aef912fc0d97784a632cc5e23f8bcb27be4966c28080c9"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.405760 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.406852 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:47.906833275 +0000 UTC m=+158.414336306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.416059 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-wslh5" podStartSLOduration=130.416039425 podStartE2EDuration="2m10.416039425s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.415898091 +0000 UTC m=+157.923401112" watchObservedRunningTime="2025-11-25 12:11:47.416039425 +0000 UTC m=+157.923542446" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.418503 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" event={"ID":"1e0bb7e2-0204-4e78-afb5-6d0ecfd59541","Type":"ContainerStarted","Data":"ffa8602d0847675dcafc991481a396b9b181676364101c0d94816f2e8cdb7883"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.424318 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.426314 4715 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5grcc container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.426373 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" podUID="1e0bb7e2-0204-4e78-afb5-6d0ecfd59541" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.445757 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf" event={"ID":"946f1f86-8ba8-4e6c-975a-b185265eb557","Type":"ContainerStarted","Data":"1f33e389bf6475ba83416dd0189e84fc180d02716a7b45266c6edab40acec837"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.472834 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p6rsr" podStartSLOduration=130.472814993 podStartE2EDuration="2m10.472814993s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.43953877 +0000 UTC m=+157.947041791" watchObservedRunningTime="2025-11-25 12:11:47.472814993 +0000 UTC m=+157.980318014" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.474157 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" event={"ID":"45106950-65c8-40eb-8dd2-c51a9b82765b","Type":"ContainerStarted","Data":"ea3b34c9020507083ca6767e7cb19cbf5ed477d42e9b310dd10d9ed4eaf31189"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.515978 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.517206 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.017176059 +0000 UTC m=+158.524679080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.534826 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-zrx9b" podStartSLOduration=130.534800737 podStartE2EDuration="2m10.534800737s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.473476551 +0000 UTC m=+157.980979582" watchObservedRunningTime="2025-11-25 12:11:47.534800737 +0000 UTC m=+158.042303758" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.543016 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tvx5q" event={"ID":"13c92ba7-a34e-4f49-b98d-032d6d8543db","Type":"ContainerStarted","Data":"f4f1b050124587aae75629252b9fb913d8484e9b2fc999cac2de970be9ff76a3"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.543062 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-tvx5q" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.549388 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-tvx5q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.549470 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tvx5q" podUID="13c92ba7-a34e-4f49-b98d-032d6d8543db" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.561424 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-8s95n" podStartSLOduration=130.561402761 podStartE2EDuration="2m10.561402761s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.559426425 +0000 UTC m=+158.066929446" watchObservedRunningTime="2025-11-25 12:11:47.561402761 +0000 UTC m=+158.068905792" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.609058 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-krnvf" podStartSLOduration=130.60903727 podStartE2EDuration="2m10.60903727s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.603556605 +0000 UTC m=+158.111059626" watchObservedRunningTime="2025-11-25 12:11:47.60903727 +0000 UTC m=+158.116540291" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.625112 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.626685 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.126665759 +0000 UTC m=+158.634168790 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.650248 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" podStartSLOduration=130.650171295 podStartE2EDuration="2m10.650171295s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.641253081 +0000 UTC m=+158.148756122" watchObservedRunningTime="2025-11-25 12:11:47.650171295 +0000 UTC m=+158.157674316" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.650684 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4dfdh" event={"ID":"25b7b306-b4b9-4a46-bc4d-38df4f035c9f","Type":"ContainerStarted","Data":"012488f685ad681c894a841043738bb5fda24086505e8a6a7c4a0d8975ec07fd"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.665474 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-tvx5q" podStartSLOduration=131.665454977 podStartE2EDuration="2m11.665454977s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.664907962 +0000 UTC m=+158.172411003" watchObservedRunningTime="2025-11-25 12:11:47.665454977 +0000 UTC m=+158.172957998" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.676951 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:47 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:47 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:47 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.677034 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.678648 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" event={"ID":"0c43525e-7149-48b7-aa59-94c4784e358b","Type":"ContainerStarted","Data":"34bcd23a7771bc6cfac73717c0afe6741e69245b9a0337d9cfb35b8ca6ad3867"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.705331 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-g5w9q" event={"ID":"b75c7c74-8c79-4f30-af10-955c80ac8abb","Type":"ContainerStarted","Data":"3bb33b8fe7a1057318bdb96c851e749830b890444f47f0c2c6002aafbb41e8a1"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.715967 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-4dfdh" podStartSLOduration=131.715942396 podStartE2EDuration="2m11.715942396s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.705730687 +0000 UTC m=+158.213233708" watchObservedRunningTime="2025-11-25 12:11:47.715942396 +0000 UTC m=+158.223445417" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.723212 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" event={"ID":"deaac6b3-5004-4f5b-a7b1-26c51247fb9d","Type":"ContainerStarted","Data":"c5bacd9c430e44ea184695b7ee7e97b4da8312e36f8b791fa870c4c23033f6f0"} Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.726817 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.727734 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.22771634 +0000 UTC m=+158.735219361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.732329 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-g5w9q" podStartSLOduration=6.73230392 podStartE2EDuration="6.73230392s" podCreationTimestamp="2025-11-25 12:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.730391005 +0000 UTC m=+158.237894036" watchObservedRunningTime="2025-11-25 12:11:47.73230392 +0000 UTC m=+158.239806941" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.735875 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.748820 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.772991 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" podStartSLOduration=130.77296148 podStartE2EDuration="2m10.77296148s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:47.765762447 +0000 UTC m=+158.273265468" watchObservedRunningTime="2025-11-25 12:11:47.77296148 +0000 UTC m=+158.280464501" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.828004 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.834927 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.334901645 +0000 UTC m=+158.842404676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.836428 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.844068 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.344052204 +0000 UTC m=+158.851555225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.937626 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:47 crc kubenswrapper[4715]: E1125 12:11:47.938069 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.438051875 +0000 UTC m=+158.945554896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.946651 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.946725 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:11:47 crc kubenswrapper[4715]: I1125 12:11:47.970781 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.039110 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.039487 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.539475587 +0000 UTC m=+159.046978608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.140068 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.140332 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.640302122 +0000 UTC m=+159.147805153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.140617 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.141028 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.641016932 +0000 UTC m=+159.148520013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.242815 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.242929 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.742908006 +0000 UTC m=+159.250411027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.243250 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.243689 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.743669898 +0000 UTC m=+159.251172919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.267467 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.267888 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.344285 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.844264656 +0000 UTC m=+159.351767677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.344322 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.344585 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.344915 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.844898545 +0000 UTC m=+159.352401566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.446763 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.447132 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:48.947114318 +0000 UTC m=+159.454617339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.548932 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.549382 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.049367974 +0000 UTC m=+159.556870995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.650635 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.650786 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.150760934 +0000 UTC m=+159.658263955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.651292 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.651749 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.151729621 +0000 UTC m=+159.659232642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.665099 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:48 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:48 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:48 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.665166 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.735574 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ttlfs" event={"ID":"deaac6b3-5004-4f5b-a7b1-26c51247fb9d","Type":"ContainerStarted","Data":"1d8b135d2178762e03f32300229d5df0bd73fe575e50eae30e50c8d924a888cc"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.738397 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-g5w9q" event={"ID":"b75c7c74-8c79-4f30-af10-955c80ac8abb","Type":"ContainerStarted","Data":"2f941573afad019ba67e9cf7641b7ec06b0a175b4cc87ae9d9696c2a0a75d888"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.753931 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" event={"ID":"6467e871-dd98-4bcf-bf7c-53726589346b","Type":"ContainerStarted","Data":"edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.754849 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.755239 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.755922 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.255902671 +0000 UTC m=+159.763405692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.756038 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2rstj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.756478 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" podUID="6467e871-dd98-4bcf-bf7c-53726589346b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.759438 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bx96c" event={"ID":"fa028433-d9fc-4729-bdb0-65cd9a4695e5","Type":"ContainerStarted","Data":"70ef3e678a9364344772dd1ae2951d1a4d076563f17297cea24f8c7c25b7ea6c"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.769088 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qmkr2" event={"ID":"b5c80b59-9196-4573-ac01-eb88a738fa25","Type":"ContainerStarted","Data":"d82b9a27ae4c627953c70a6f6a3261186ef805225600c5f7d07fcceb4a234033"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.785214 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" event={"ID":"5cf156ab-e8de-4545-b432-dd538b3c7ce4","Type":"ContainerStarted","Data":"77ed547c56cbf3192cd7b39e0878ba98b7a4e53bbf4a14a449d1d75c9ff77504"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.785258 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" event={"ID":"5cf156ab-e8de-4545-b432-dd538b3c7ce4","Type":"ContainerStarted","Data":"af5f6debff8f4644f648f14bd2e83e8bac975e2bd29f542bce585fc38373d33b"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.789851 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" event={"ID":"0c43525e-7149-48b7-aa59-94c4784e358b","Type":"ContainerStarted","Data":"ce4be955f1a1da3325207434f64bae967f1dbdf0aad2f2bd810cec4dd69c4b38"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.789884 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" event={"ID":"0c43525e-7149-48b7-aa59-94c4784e358b","Type":"ContainerStarted","Data":"450dea917fac5add476adf378e2ac5fc3f17b2f526a43105153c2f36f821414a"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.790315 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.799314 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" podStartSLOduration=131.79929874 podStartE2EDuration="2m11.79929874s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:48.798337382 +0000 UTC m=+159.305840413" watchObservedRunningTime="2025-11-25 12:11:48.79929874 +0000 UTC m=+159.306801761" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.801059 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-p6m6d" event={"ID":"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf","Type":"ContainerStarted","Data":"7bcaa73eb1bb23311abcde6d8e9fb070bbd9eca4e4ca729e3403ba7047863f33"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.801101 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-p6m6d" event={"ID":"4d62ff1d-b062-4dfc-ba55-eae6dd68feaf","Type":"ContainerStarted","Data":"cad3d7022f82bfdbfb01b55c1a1c61fa619266e15c8cbb927ac769b1f52059d8"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.801924 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.816421 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" event={"ID":"19cbbcfc-8187-4b59-a1e3-3a33e64e37c4","Type":"ContainerStarted","Data":"7dfc7b31bc182ef3baef9d86a4fd64624e758c3a79d625afc42ddaf36eec4c12"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.822699 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5bct" event={"ID":"f5204538-5658-4148-8a88-7590e51114c5","Type":"ContainerStarted","Data":"f1c3a569d877e0f60690aac23e6b2bca2390532c5db08c39edf9ba81e68d2ad0"} Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.826443 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-tvx5q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.826486 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tvx5q" podUID="13c92ba7-a34e-4f49-b98d-032d6d8543db" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.831409 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" podStartSLOduration=131.831390208 podStartE2EDuration="2m11.831390208s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:48.831329016 +0000 UTC m=+159.338832037" watchObservedRunningTime="2025-11-25 12:11:48.831390208 +0000 UTC m=+159.338893239" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.842835 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5grcc" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.851499 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-clmh7" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.866430 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.875961 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.3759478 +0000 UTC m=+159.883450821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.877994 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" podStartSLOduration=132.877978967 podStartE2EDuration="2m12.877978967s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:48.874484559 +0000 UTC m=+159.381987580" watchObservedRunningTime="2025-11-25 12:11:48.877978967 +0000 UTC m=+159.385481988" Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.968702 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.969058 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.469036135 +0000 UTC m=+159.976539146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.969615 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:48 crc kubenswrapper[4715]: E1125 12:11:48.975326 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.475309693 +0000 UTC m=+159.982812714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:48 crc kubenswrapper[4715]: I1125 12:11:48.990612 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-f6n6v" podStartSLOduration=131.990586255 podStartE2EDuration="2m11.990586255s" podCreationTimestamp="2025-11-25 12:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:48.975116668 +0000 UTC m=+159.482619679" watchObservedRunningTime="2025-11-25 12:11:48.990586255 +0000 UTC m=+159.498089286" Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.071296 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.071596 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.571572878 +0000 UTC m=+160.079075899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.079152 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-p6m6d" podStartSLOduration=8.079138713 podStartE2EDuration="8.079138713s" podCreationTimestamp="2025-11-25 12:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:49.077055274 +0000 UTC m=+159.584558285" watchObservedRunningTime="2025-11-25 12:11:49.079138713 +0000 UTC m=+159.586641734" Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.100591 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.158621 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cj9n6" Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.173227 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.173533 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.673522415 +0000 UTC m=+160.181025436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.274306 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.275915 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.775895573 +0000 UTC m=+160.283398604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.376740 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.377118 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.877104069 +0000 UTC m=+160.384607090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.477680 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.477808 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.9777887 +0000 UTC m=+160.485291721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.477934 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.478328 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:49.978320534 +0000 UTC m=+160.485823555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.579237 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.579557 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.07954293 +0000 UTC m=+160.587045951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.664153 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:49 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:49 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:49 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.664528 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.680784 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.681321 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.181303922 +0000 UTC m=+160.688806943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.781943 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.782133 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.282106776 +0000 UTC m=+160.789609797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.783117 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.783541 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.283532416 +0000 UTC m=+160.791035437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.826440 4715 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-w4bkk container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.826502 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" podUID="712c4b61-4bc2-41b0-8837-3d2d51e2a288" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.840693 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2rstj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.840975 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" podUID="6467e871-dd98-4bcf-bf7c-53726589346b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.848859 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zmbnj" Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.892897 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.893362 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.393337735 +0000 UTC m=+160.900840756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.894681 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:49 crc kubenswrapper[4715]: E1125 12:11:49.899124 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.399109528 +0000 UTC m=+160.906612549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:49 crc kubenswrapper[4715]: I1125 12:11:49.975871 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w4bkk" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.000407 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.000590 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.500560821 +0000 UTC m=+161.008063852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.001073 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.001486 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.501474097 +0000 UTC m=+161.008977118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.102160 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.102356 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.602323482 +0000 UTC m=+161.109826513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.102425 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.102723 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.602708293 +0000 UTC m=+161.110211314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.203350 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.203740 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.703718443 +0000 UTC m=+161.211221474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.210712 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zxjdm"] Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.212017 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: W1125 12:11:50.217755 4715 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.218001 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.305461 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r55px\" (UniqueName: \"kubernetes.io/projected/a3a35017-fd36-43de-83bd-8787fece173a-kube-api-access-r55px\") pod \"community-operators-zxjdm\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.306028 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.306215 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-utilities\") pod \"community-operators-zxjdm\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.306505 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.806485693 +0000 UTC m=+161.313988774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.306500 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-catalog-content\") pod \"community-operators-zxjdm\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.358923 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gdrl9"] Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.360333 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.366465 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zxjdm"] Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.380682 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.396680 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gdrl9"] Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.407637 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.407803 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.90776833 +0000 UTC m=+161.415271351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.408733 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-utilities\") pod \"certified-operators-gdrl9\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.408912 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r55px\" (UniqueName: \"kubernetes.io/projected/a3a35017-fd36-43de-83bd-8787fece173a-kube-api-access-r55px\") pod \"community-operators-zxjdm\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.408968 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.409062 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-catalog-content\") pod \"certified-operators-gdrl9\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.409132 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-utilities\") pod \"community-operators-zxjdm\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.409178 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjv22\" (UniqueName: \"kubernetes.io/projected/6af7e329-2107-4548-b087-77b8b60f6646-kube-api-access-vjv22\") pod \"certified-operators-gdrl9\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.409273 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-catalog-content\") pod \"community-operators-zxjdm\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.409377 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:50.909363956 +0000 UTC m=+161.416866977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.409800 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-utilities\") pod \"community-operators-zxjdm\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.409868 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-catalog-content\") pod \"community-operators-zxjdm\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.455820 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r55px\" (UniqueName: \"kubernetes.io/projected/a3a35017-fd36-43de-83bd-8787fece173a-kube-api-access-r55px\") pod \"community-operators-zxjdm\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.510922 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.511163 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-utilities\") pod \"certified-operators-gdrl9\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.511265 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-catalog-content\") pod \"certified-operators-gdrl9\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.511302 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjv22\" (UniqueName: \"kubernetes.io/projected/6af7e329-2107-4548-b087-77b8b60f6646-kube-api-access-vjv22\") pod \"certified-operators-gdrl9\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.511737 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.011717063 +0000 UTC m=+161.519220084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.537385 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l7rv8"] Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.538293 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.564927 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l7rv8"] Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.613032 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5w78\" (UniqueName: \"kubernetes.io/projected/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-kube-api-access-l5w78\") pod \"community-operators-l7rv8\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.613093 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-utilities\") pod \"community-operators-l7rv8\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.613209 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-catalog-content\") pod \"community-operators-l7rv8\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.613292 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.613667 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.113650669 +0000 UTC m=+161.621153690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.663707 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:50 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:50 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:50 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.663796 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.673397 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-utilities\") pod \"certified-operators-gdrl9\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.674237 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-catalog-content\") pod \"certified-operators-gdrl9\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.676768 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjv22\" (UniqueName: \"kubernetes.io/projected/6af7e329-2107-4548-b087-77b8b60f6646-kube-api-access-vjv22\") pod \"certified-operators-gdrl9\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.677088 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.714593 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.714922 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5w78\" (UniqueName: \"kubernetes.io/projected/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-kube-api-access-l5w78\") pod \"community-operators-l7rv8\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.714961 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-utilities\") pod \"community-operators-l7rv8\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.715036 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-catalog-content\") pod \"community-operators-l7rv8\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.715607 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-catalog-content\") pod \"community-operators-l7rv8\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.715700 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.215679958 +0000 UTC m=+161.723182989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.716278 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-utilities\") pod \"community-operators-l7rv8\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.738418 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5w78\" (UniqueName: \"kubernetes.io/projected/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-kube-api-access-l5w78\") pod \"community-operators-l7rv8\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.757618 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z2nhl"] Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.758842 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.760832 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2nhl"] Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.816027 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.816316 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-catalog-content\") pod \"certified-operators-z2nhl\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.816357 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-utilities\") pod \"certified-operators-z2nhl\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.816375 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdbfx\" (UniqueName: \"kubernetes.io/projected/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-kube-api-access-jdbfx\") pod \"certified-operators-z2nhl\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.816702 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.316689528 +0000 UTC m=+161.824192539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.890957 4715 generic.go:334] "Generic (PLEG): container finished" podID="829f6cba-1de5-4a2c-9f72-e24fe514caa9" containerID="4210d8fd6ab3f8692a875f90e733091d229bb410a2edc91256c094a426cf95f0" exitCode=0 Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.891995 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" event={"ID":"829f6cba-1de5-4a2c-9f72-e24fe514caa9","Type":"ContainerDied","Data":"4210d8fd6ab3f8692a875f90e733091d229bb410a2edc91256c094a426cf95f0"} Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.925780 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.926265 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-catalog-content\") pod \"certified-operators-z2nhl\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.926327 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-utilities\") pod \"certified-operators-z2nhl\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.926351 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdbfx\" (UniqueName: \"kubernetes.io/projected/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-kube-api-access-jdbfx\") pod \"certified-operators-z2nhl\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:50 crc kubenswrapper[4715]: E1125 12:11:50.926770 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.426750754 +0000 UTC m=+161.934253785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.927309 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-catalog-content\") pod \"certified-operators-z2nhl\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.927594 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-utilities\") pod \"certified-operators-z2nhl\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:50 crc kubenswrapper[4715]: I1125 12:11:50.961494 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdbfx\" (UniqueName: \"kubernetes.io/projected/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-kube-api-access-jdbfx\") pod \"certified-operators-z2nhl\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.021197 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gdrl9"] Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.029251 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.030296 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.530279475 +0000 UTC m=+162.037782496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: W1125 12:11:51.046524 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6af7e329_2107_4548_b087_77b8b60f6646.slice/crio-6b3811d4d2322fb37501b2c0fbf2084e3b1fa3365e5eb3e0d32ec00fd8bbd140 WatchSource:0}: Error finding container 6b3811d4d2322fb37501b2c0fbf2084e3b1fa3365e5eb3e0d32ec00fd8bbd140: Status 404 returned error can't find the container with id 6b3811d4d2322fb37501b2c0fbf2084e3b1fa3365e5eb3e0d32ec00fd8bbd140 Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.084806 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.131465 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.131946 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.631924343 +0000 UTC m=+162.139427364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.233436 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.234174 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.734158078 +0000 UTC m=+162.241661109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.335113 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.335263 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.83524678 +0000 UTC m=+162.342749801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.335439 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.335724 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.835715233 +0000 UTC m=+162.343218264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.374664 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2nhl"] Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.375417 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.380336 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.384641 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:11:51 crc kubenswrapper[4715]: W1125 12:11:51.406874 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc4608d5_28b1_4b1a_94da_72f4f7bb7907.slice/crio-bf24d5bd240f6f5801e5f30a2d202e40e38ef3c82c7e390e975576afa34f11bf WatchSource:0}: Error finding container bf24d5bd240f6f5801e5f30a2d202e40e38ef3c82c7e390e975576afa34f11bf: Status 404 returned error can't find the container with id bf24d5bd240f6f5801e5f30a2d202e40e38ef3c82c7e390e975576afa34f11bf Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.436841 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.437028 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.93700396 +0000 UTC m=+162.444506981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.437254 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.437599 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:51.937591098 +0000 UTC m=+162.445094119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.538329 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.538466 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.038444323 +0000 UTC m=+162.545947344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.539065 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.539474 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.039456421 +0000 UTC m=+162.546959442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.609392 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l7rv8"] Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.641101 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.641332 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.141301755 +0000 UTC m=+162.648804776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.641643 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.642047 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.142035936 +0000 UTC m=+162.649538957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.658004 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zxjdm"] Nov 25 12:11:51 crc kubenswrapper[4715]: W1125 12:11:51.664211 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3a35017_fd36_43de_83bd_8787fece173a.slice/crio-457d5b985d3acc4616241e0b8ff98a65e1f200fe38983a1a15f401e97aa34a97 WatchSource:0}: Error finding container 457d5b985d3acc4616241e0b8ff98a65e1f200fe38983a1a15f401e97aa34a97: Status 404 returned error can't find the container with id 457d5b985d3acc4616241e0b8ff98a65e1f200fe38983a1a15f401e97aa34a97 Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.664973 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:51 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:51 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:51 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.665029 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.742404 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.742574 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.242551301 +0000 UTC m=+162.750054332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.742758 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.743089 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.243081547 +0000 UTC m=+162.750584568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.806236 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6af7e329_2107_4548_b087_77b8b60f6646.slice/crio-conmon-4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b.scope\": RecentStats: unable to find data in memory cache]" Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.843554 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.843955 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.343936162 +0000 UTC m=+162.851439183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.899138 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5bct" event={"ID":"f5204538-5658-4148-8a88-7590e51114c5","Type":"ContainerStarted","Data":"599fa26bb4fa78c3e0f119ebc9795337149ab6bedb67c449c8e2ba094186bea5"} Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.900913 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l7rv8" event={"ID":"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3","Type":"ContainerStarted","Data":"d75834ab177f8e867abc49607b9dd8c591e1b452b004f5a3753647c2faba3dcc"} Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.903111 4715 generic.go:334] "Generic (PLEG): container finished" podID="6af7e329-2107-4548-b087-77b8b60f6646" containerID="4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b" exitCode=0 Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.903219 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdrl9" event={"ID":"6af7e329-2107-4548-b087-77b8b60f6646","Type":"ContainerDied","Data":"4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b"} Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.903257 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdrl9" event={"ID":"6af7e329-2107-4548-b087-77b8b60f6646","Type":"ContainerStarted","Data":"6b3811d4d2322fb37501b2c0fbf2084e3b1fa3365e5eb3e0d32ec00fd8bbd140"} Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.906479 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxjdm" event={"ID":"a3a35017-fd36-43de-83bd-8787fece173a","Type":"ContainerStarted","Data":"457d5b985d3acc4616241e0b8ff98a65e1f200fe38983a1a15f401e97aa34a97"} Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.907803 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.908504 4715 generic.go:334] "Generic (PLEG): container finished" podID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerID="f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1" exitCode=0 Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.908688 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2nhl" event={"ID":"cc4608d5-28b1-4b1a-94da-72f4f7bb7907","Type":"ContainerDied","Data":"f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1"} Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.908757 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2nhl" event={"ID":"cc4608d5-28b1-4b1a-94da-72f4f7bb7907","Type":"ContainerStarted","Data":"bf24d5bd240f6f5801e5f30a2d202e40e38ef3c82c7e390e975576afa34f11bf"} Nov 25 12:11:51 crc kubenswrapper[4715]: I1125 12:11:51.945046 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:51 crc kubenswrapper[4715]: E1125 12:11:51.945492 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.445472397 +0000 UTC m=+162.952975418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.046519 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.048238 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.548138533 +0000 UTC m=+163.055641554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.092895 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.125471 4715 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.129556 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rgjtn"] Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.129804 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829f6cba-1de5-4a2c-9f72-e24fe514caa9" containerName="collect-profiles" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.129823 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="829f6cba-1de5-4a2c-9f72-e24fe514caa9" containerName="collect-profiles" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.129949 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="829f6cba-1de5-4a2c-9f72-e24fe514caa9" containerName="collect-profiles" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.130833 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.134534 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.137880 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rgjtn"] Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.148275 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4b7q\" (UniqueName: \"kubernetes.io/projected/829f6cba-1de5-4a2c-9f72-e24fe514caa9-kube-api-access-b4b7q\") pod \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.148539 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/829f6cba-1de5-4a2c-9f72-e24fe514caa9-config-volume\") pod \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.148568 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/829f6cba-1de5-4a2c-9f72-e24fe514caa9-secret-volume\") pod \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\" (UID: \"829f6cba-1de5-4a2c-9f72-e24fe514caa9\") " Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.148785 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.149155 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/829f6cba-1de5-4a2c-9f72-e24fe514caa9-config-volume" (OuterVolumeSpecName: "config-volume") pod "829f6cba-1de5-4a2c-9f72-e24fe514caa9" (UID: "829f6cba-1de5-4a2c-9f72-e24fe514caa9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.149203 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.649155914 +0000 UTC m=+163.156658935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.153847 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829f6cba-1de5-4a2c-9f72-e24fe514caa9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "829f6cba-1de5-4a2c-9f72-e24fe514caa9" (UID: "829f6cba-1de5-4a2c-9f72-e24fe514caa9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.158098 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829f6cba-1de5-4a2c-9f72-e24fe514caa9-kube-api-access-b4b7q" (OuterVolumeSpecName: "kube-api-access-b4b7q") pod "829f6cba-1de5-4a2c-9f72-e24fe514caa9" (UID: "829f6cba-1de5-4a2c-9f72-e24fe514caa9"). InnerVolumeSpecName "kube-api-access-b4b7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.249832 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.250208 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.750138793 +0000 UTC m=+163.257641814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.250435 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4k5w\" (UniqueName: \"kubernetes.io/projected/b3d2d042-e214-44ff-9c25-9d894bab99f3-kube-api-access-b4k5w\") pod \"redhat-marketplace-rgjtn\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.250620 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-catalog-content\") pod \"redhat-marketplace-rgjtn\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.250780 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.250931 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-utilities\") pod \"redhat-marketplace-rgjtn\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.251084 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.751069059 +0000 UTC m=+163.258572080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.251243 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/829f6cba-1de5-4a2c-9f72-e24fe514caa9-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.251330 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/829f6cba-1de5-4a2c-9f72-e24fe514caa9-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.251404 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4b7q\" (UniqueName: \"kubernetes.io/projected/829f6cba-1de5-4a2c-9f72-e24fe514caa9-kube-api-access-b4b7q\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.352477 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.352763 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4k5w\" (UniqueName: \"kubernetes.io/projected/b3d2d042-e214-44ff-9c25-9d894bab99f3-kube-api-access-b4k5w\") pod \"redhat-marketplace-rgjtn\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.352833 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-catalog-content\") pod \"redhat-marketplace-rgjtn\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.352903 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-utilities\") pod \"redhat-marketplace-rgjtn\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.353385 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-utilities\") pod \"redhat-marketplace-rgjtn\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.353479 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.853460328 +0000 UTC m=+163.360963349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.353984 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-catalog-content\") pod \"redhat-marketplace-rgjtn\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.375857 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4k5w\" (UniqueName: \"kubernetes.io/projected/b3d2d042-e214-44ff-9c25-9d894bab99f3-kube-api-access-b4k5w\") pod \"redhat-marketplace-rgjtn\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.451295 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.453977 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.454496 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:52.954479179 +0000 UTC m=+163.461982200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.524657 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mwc6v"] Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.525917 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.544963 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwc6v"] Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.555426 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.555741 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:53.055726195 +0000 UTC m=+163.563229216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.657100 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wx96\" (UniqueName: \"kubernetes.io/projected/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-kube-api-access-6wx96\") pod \"redhat-marketplace-mwc6v\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.657201 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.657264 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-catalog-content\") pod \"redhat-marketplace-mwc6v\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.657701 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:53.15763276 +0000 UTC m=+163.665135861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.658308 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-utilities\") pod \"redhat-marketplace-mwc6v\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.668793 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:52 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:52 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:52 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.668854 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.669084 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rgjtn"] Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.764597 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.765412 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wx96\" (UniqueName: \"kubernetes.io/projected/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-kube-api-access-6wx96\") pod \"redhat-marketplace-mwc6v\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.765487 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-catalog-content\") pod \"redhat-marketplace-mwc6v\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.765536 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-utilities\") pod \"redhat-marketplace-mwc6v\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.766290 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:53.266266866 +0000 UTC m=+163.773769897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.766369 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-utilities\") pod \"redhat-marketplace-mwc6v\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.766569 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-catalog-content\") pod \"redhat-marketplace-mwc6v\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.789045 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wx96\" (UniqueName: \"kubernetes.io/projected/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-kube-api-access-6wx96\") pod \"redhat-marketplace-mwc6v\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.849207 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.867131 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.868044 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:53.368026487 +0000 UTC m=+163.875529508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.945654 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5bct" event={"ID":"f5204538-5658-4148-8a88-7590e51114c5","Type":"ContainerStarted","Data":"23201f99bc05831adf49c2af4bbab30199ec91c208bbc95174daef2358bad059"} Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.945984 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5bct" event={"ID":"f5204538-5658-4148-8a88-7590e51114c5","Type":"ContainerStarted","Data":"5bdfbf85b208f2dca5dc82d3d8151ca07f3bb99fbaefb523bd23f5448bd49a13"} Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.952954 4715 generic.go:334] "Generic (PLEG): container finished" podID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerID="634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d" exitCode=0 Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.962242 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rgjtn" event={"ID":"b3d2d042-e214-44ff-9c25-9d894bab99f3","Type":"ContainerDied","Data":"634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d"} Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.962326 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rgjtn" event={"ID":"b3d2d042-e214-44ff-9c25-9d894bab99f3","Type":"ContainerStarted","Data":"a6c5a6bbff0b28e073b9b5f88d096caae3451daf427ec89c362601763dcb243b"} Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.965488 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.965517 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-zr2x7" event={"ID":"829f6cba-1de5-4a2c-9f72-e24fe514caa9","Type":"ContainerDied","Data":"cd2a32af04e2aeb42bad1a446ec3fe2309e1bee6e9162ed05ec767b91c092131"} Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.965588 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd2a32af04e2aeb42bad1a446ec3fe2309e1bee6e9162ed05ec767b91c092131" Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.971149 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:52 crc kubenswrapper[4715]: E1125 12:11:52.971625 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 12:11:53.47160143 +0000 UTC m=+163.979104451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.983612 4715 generic.go:334] "Generic (PLEG): container finished" podID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerID="b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d" exitCode=0 Nov 25 12:11:52 crc kubenswrapper[4715]: I1125 12:11:52.983681 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l7rv8" event={"ID":"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3","Type":"ContainerDied","Data":"b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d"} Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.000109 4715 generic.go:334] "Generic (PLEG): container finished" podID="a3a35017-fd36-43de-83bd-8787fece173a" containerID="89e1f5578d926e6633f83eb4ff0a5f6e8481a67d17a19f11565ec7fa60fe867b" exitCode=0 Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.000221 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxjdm" event={"ID":"a3a35017-fd36-43de-83bd-8787fece173a","Type":"ContainerDied","Data":"89e1f5578d926e6633f83eb4ff0a5f6e8481a67d17a19f11565ec7fa60fe867b"} Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.015609 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-n5bct" podStartSLOduration=12.015577285 podStartE2EDuration="12.015577285s" podCreationTimestamp="2025-11-25 12:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:52.978261799 +0000 UTC m=+163.485764840" watchObservedRunningTime="2025-11-25 12:11:53.015577285 +0000 UTC m=+163.523080326" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.065548 4715 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T12:11:52.125504204Z","Handler":null,"Name":""} Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.084124 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:53 crc kubenswrapper[4715]: E1125 12:11:53.084884 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 12:11:53.584868097 +0000 UTC m=+164.092371118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kbdhr" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.087285 4715 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.087331 4715 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.127315 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwc6v"] Nov 25 12:11:53 crc kubenswrapper[4715]: W1125 12:11:53.163317 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55c59d7a_0cdb_48e7_a315_f2a3fdeee4df.slice/crio-1eba42764eb2367c0e2203b65b900874dc252827250bac4d71afb8b9eba1a9be WatchSource:0}: Error finding container 1eba42764eb2367c0e2203b65b900874dc252827250bac4d71afb8b9eba1a9be: Status 404 returned error can't find the container with id 1eba42764eb2367c0e2203b65b900874dc252827250bac4d71afb8b9eba1a9be Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.185744 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.206025 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.291094 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.302321 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.303499 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.306670 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.317154 4715 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.317214 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.317690 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.323837 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.354326 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2bpdr"] Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.355366 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2bpdr"] Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.355509 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.357023 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.377962 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kbdhr\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.393700 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48ead74-9fba-4462-88f9-e1207f28d880-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e48ead74-9fba-4462-88f9-e1207f28d880\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.393762 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48ead74-9fba-4462-88f9-e1207f28d880-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e48ead74-9fba-4462-88f9-e1207f28d880\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.499494 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-utilities\") pod \"redhat-operators-2bpdr\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.499592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48ead74-9fba-4462-88f9-e1207f28d880-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e48ead74-9fba-4462-88f9-e1207f28d880\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.499665 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75wb8\" (UniqueName: \"kubernetes.io/projected/6da3d717-99fa-4202-909e-45db02ba3385-kube-api-access-75wb8\") pod \"redhat-operators-2bpdr\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.499726 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48ead74-9fba-4462-88f9-e1207f28d880-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e48ead74-9fba-4462-88f9-e1207f28d880\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.499755 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-catalog-content\") pod \"redhat-operators-2bpdr\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.500003 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48ead74-9fba-4462-88f9-e1207f28d880-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e48ead74-9fba-4462-88f9-e1207f28d880\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.527379 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48ead74-9fba-4462-88f9-e1207f28d880-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e48ead74-9fba-4462-88f9-e1207f28d880\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.569097 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.571604 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.577076 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.577334 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.578449 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.587689 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.589674 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.589833 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.598339 4715 patch_prober.go:28] interesting pod/apiserver-76f77b778f-dl4rf container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]log ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]etcd ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/generic-apiserver-start-informers ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/max-in-flight-filter ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 25 12:11:53 crc kubenswrapper[4715]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 25 12:11:53 crc kubenswrapper[4715]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/project.openshift.io-projectcache ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/openshift.io-startinformers ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 25 12:11:53 crc kubenswrapper[4715]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 25 12:11:53 crc kubenswrapper[4715]: livez check failed Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.598450 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" podUID="5cf156ab-e8de-4545-b432-dd538b3c7ce4" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.601300 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-catalog-content\") pod \"redhat-operators-2bpdr\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.601451 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-utilities\") pod \"redhat-operators-2bpdr\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.601507 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75wb8\" (UniqueName: \"kubernetes.io/projected/6da3d717-99fa-4202-909e-45db02ba3385-kube-api-access-75wb8\") pod \"redhat-operators-2bpdr\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.602064 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-catalog-content\") pod \"redhat-operators-2bpdr\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.602910 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-utilities\") pod \"redhat-operators-2bpdr\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.621394 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75wb8\" (UniqueName: \"kubernetes.io/projected/6da3d717-99fa-4202-909e-45db02ba3385-kube-api-access-75wb8\") pod \"redhat-operators-2bpdr\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.644978 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.661082 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.667269 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:53 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:53 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:53 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.667331 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.673840 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.703164 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b05a676-12ac-429d-9e28-f5950f29f37b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4b05a676-12ac-429d-9e28-f5950f29f37b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.703265 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b05a676-12ac-429d-9e28-f5950f29f37b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4b05a676-12ac-429d-9e28-f5950f29f37b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.733499 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wn6k5"] Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.737235 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.744677 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wn6k5"] Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.812594 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-catalog-content\") pod \"redhat-operators-wn6k5\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.812738 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btcqn\" (UniqueName: \"kubernetes.io/projected/2b87d8f5-eda7-422d-ae7b-502725559317-kube-api-access-btcqn\") pod \"redhat-operators-wn6k5\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.812794 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-utilities\") pod \"redhat-operators-wn6k5\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.812842 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b05a676-12ac-429d-9e28-f5950f29f37b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4b05a676-12ac-429d-9e28-f5950f29f37b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.812870 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b05a676-12ac-429d-9e28-f5950f29f37b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4b05a676-12ac-429d-9e28-f5950f29f37b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.816328 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b05a676-12ac-429d-9e28-f5950f29f37b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4b05a676-12ac-429d-9e28-f5950f29f37b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.835202 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-tvx5q container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.835260 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-tvx5q" podUID="13c92ba7-a34e-4f49-b98d-032d6d8543db" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.835083 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-tvx5q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.835568 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tvx5q" podUID="13c92ba7-a34e-4f49-b98d-032d6d8543db" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.841210 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.843379 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.854465 4715 patch_prober.go:28] interesting pod/console-f9d7485db-4dfdh container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.854579 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-4dfdh" podUID="25b7b306-b4b9-4a46-bc4d-38df4f035c9f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.854943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b05a676-12ac-429d-9e28-f5950f29f37b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4b05a676-12ac-429d-9e28-f5950f29f37b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.856480 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kbdhr"] Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.895008 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.914366 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-catalog-content\") pod \"redhat-operators-wn6k5\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.914458 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btcqn\" (UniqueName: \"kubernetes.io/projected/2b87d8f5-eda7-422d-ae7b-502725559317-kube-api-access-btcqn\") pod \"redhat-operators-wn6k5\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.914501 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-utilities\") pod \"redhat-operators-wn6k5\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.915380 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-utilities\") pod \"redhat-operators-wn6k5\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.915685 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-catalog-content\") pod \"redhat-operators-wn6k5\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.927915 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:11:53 crc kubenswrapper[4715]: I1125 12:11:53.953662 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btcqn\" (UniqueName: \"kubernetes.io/projected/2b87d8f5-eda7-422d-ae7b-502725559317-kube-api-access-btcqn\") pod \"redhat-operators-wn6k5\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.048774 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" event={"ID":"3088de81-03d3-483f-b6c1-ddf0a6e30c34","Type":"ContainerStarted","Data":"ab54f30daef82c55ac5e236b047e9ccb17a70f55e0b5c9e87aa739823435f9f3"} Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.051316 4715 generic.go:334] "Generic (PLEG): container finished" podID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerID="2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf" exitCode=0 Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.052901 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwc6v" event={"ID":"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df","Type":"ContainerDied","Data":"2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf"} Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.052925 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwc6v" event={"ID":"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df","Type":"ContainerStarted","Data":"1eba42764eb2367c0e2203b65b900874dc252827250bac4d71afb8b9eba1a9be"} Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.075630 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.172378 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2bpdr"] Nov 25 12:11:54 crc kubenswrapper[4715]: W1125 12:11:54.199640 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6da3d717_99fa_4202_909e_45db02ba3385.slice/crio-0ec3964531fe1b70f1974e66a2d6f6c654016b8ea416c059ff4baf4b17097c98 WatchSource:0}: Error finding container 0ec3964531fe1b70f1974e66a2d6f6c654016b8ea416c059ff4baf4b17097c98: Status 404 returned error can't find the container with id 0ec3964531fe1b70f1974e66a2d6f6c654016b8ea416c059ff4baf4b17097c98 Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.279469 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.367843 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.552014 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wn6k5"] Nov 25 12:11:54 crc kubenswrapper[4715]: W1125 12:11:54.621497 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b87d8f5_eda7_422d_ae7b_502725559317.slice/crio-baea65f3b3ca21551daf20ddd333918d5e445e57dc763ee1117b907903e9aab0 WatchSource:0}: Error finding container baea65f3b3ca21551daf20ddd333918d5e445e57dc763ee1117b907903e9aab0: Status 404 returned error can't find the container with id baea65f3b3ca21551daf20ddd333918d5e445e57dc763ee1117b907903e9aab0 Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.670348 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:54 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:54 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:54 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.670437 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:54 crc kubenswrapper[4715]: I1125 12:11:54.756927 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.119792 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bpdr" event={"ID":"6da3d717-99fa-4202-909e-45db02ba3385","Type":"ContainerDied","Data":"033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8"} Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.120155 4715 generic.go:334] "Generic (PLEG): container finished" podID="6da3d717-99fa-4202-909e-45db02ba3385" containerID="033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8" exitCode=0 Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.120539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bpdr" event={"ID":"6da3d717-99fa-4202-909e-45db02ba3385","Type":"ContainerStarted","Data":"0ec3964531fe1b70f1974e66a2d6f6c654016b8ea416c059ff4baf4b17097c98"} Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.125071 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" event={"ID":"3088de81-03d3-483f-b6c1-ddf0a6e30c34","Type":"ContainerStarted","Data":"d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079"} Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.125170 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.133462 4715 generic.go:334] "Generic (PLEG): container finished" podID="2b87d8f5-eda7-422d-ae7b-502725559317" containerID="31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263" exitCode=0 Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.133626 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wn6k5" event={"ID":"2b87d8f5-eda7-422d-ae7b-502725559317","Type":"ContainerDied","Data":"31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263"} Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.133683 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wn6k5" event={"ID":"2b87d8f5-eda7-422d-ae7b-502725559317","Type":"ContainerStarted","Data":"baea65f3b3ca21551daf20ddd333918d5e445e57dc763ee1117b907903e9aab0"} Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.147310 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e48ead74-9fba-4462-88f9-e1207f28d880","Type":"ContainerStarted","Data":"ed7cc62c7bd1ebe99011684e0386b6fc4644744a43b4c0e5251cbdbf41ee67d2"} Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.150313 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4b05a676-12ac-429d-9e28-f5950f29f37b","Type":"ContainerStarted","Data":"0cd250a4823a894ac06f579c50323cea747a8b2e83804f1fec657d2ba0c2b0d7"} Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.167057 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" podStartSLOduration=139.167034999 podStartE2EDuration="2m19.167034999s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:55.164409685 +0000 UTC m=+165.671912706" watchObservedRunningTime="2025-11-25 12:11:55.167034999 +0000 UTC m=+165.674538020" Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.212082 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.212057254 podStartE2EDuration="2.212057254s" podCreationTimestamp="2025-11-25 12:11:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:11:55.20554557 +0000 UTC m=+165.713048611" watchObservedRunningTime="2025-11-25 12:11:55.212057254 +0000 UTC m=+165.719560285" Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.665103 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:55 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:55 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:55 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:55 crc kubenswrapper[4715]: I1125 12:11:55.665160 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:56 crc kubenswrapper[4715]: I1125 12:11:56.157892 4715 generic.go:334] "Generic (PLEG): container finished" podID="e48ead74-9fba-4462-88f9-e1207f28d880" containerID="0e0b453ead5d6405a2944ee45d3f502abb479caffda6ba32a7b95ac8f32c75c3" exitCode=0 Nov 25 12:11:56 crc kubenswrapper[4715]: I1125 12:11:56.157999 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e48ead74-9fba-4462-88f9-e1207f28d880","Type":"ContainerDied","Data":"0e0b453ead5d6405a2944ee45d3f502abb479caffda6ba32a7b95ac8f32c75c3"} Nov 25 12:11:56 crc kubenswrapper[4715]: I1125 12:11:56.160646 4715 generic.go:334] "Generic (PLEG): container finished" podID="4b05a676-12ac-429d-9e28-f5950f29f37b" containerID="4654f1fa81a134637a38d51cea69b91f8c61c508c89199e5a9235bc671d87b86" exitCode=0 Nov 25 12:11:56 crc kubenswrapper[4715]: I1125 12:11:56.160754 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4b05a676-12ac-429d-9e28-f5950f29f37b","Type":"ContainerDied","Data":"4654f1fa81a134637a38d51cea69b91f8c61c508c89199e5a9235bc671d87b86"} Nov 25 12:11:56 crc kubenswrapper[4715]: I1125 12:11:56.664574 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:56 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:56 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:56 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:56 crc kubenswrapper[4715]: I1125 12:11:56.665332 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.605677 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.607952 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.663082 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:57 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:57 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:57 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.663141 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.690270 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48ead74-9fba-4462-88f9-e1207f28d880-kubelet-dir\") pod \"e48ead74-9fba-4462-88f9-e1207f28d880\" (UID: \"e48ead74-9fba-4462-88f9-e1207f28d880\") " Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.690335 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b05a676-12ac-429d-9e28-f5950f29f37b-kube-api-access\") pod \"4b05a676-12ac-429d-9e28-f5950f29f37b\" (UID: \"4b05a676-12ac-429d-9e28-f5950f29f37b\") " Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.690374 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48ead74-9fba-4462-88f9-e1207f28d880-kube-api-access\") pod \"e48ead74-9fba-4462-88f9-e1207f28d880\" (UID: \"e48ead74-9fba-4462-88f9-e1207f28d880\") " Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.690426 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e48ead74-9fba-4462-88f9-e1207f28d880-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e48ead74-9fba-4462-88f9-e1207f28d880" (UID: "e48ead74-9fba-4462-88f9-e1207f28d880"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.690454 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b05a676-12ac-429d-9e28-f5950f29f37b-kubelet-dir\") pod \"4b05a676-12ac-429d-9e28-f5950f29f37b\" (UID: \"4b05a676-12ac-429d-9e28-f5950f29f37b\") " Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.690764 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e48ead74-9fba-4462-88f9-e1207f28d880-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.690809 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b05a676-12ac-429d-9e28-f5950f29f37b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4b05a676-12ac-429d-9e28-f5950f29f37b" (UID: "4b05a676-12ac-429d-9e28-f5950f29f37b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.695786 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e48ead74-9fba-4462-88f9-e1207f28d880-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e48ead74-9fba-4462-88f9-e1207f28d880" (UID: "e48ead74-9fba-4462-88f9-e1207f28d880"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.696012 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b05a676-12ac-429d-9e28-f5950f29f37b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4b05a676-12ac-429d-9e28-f5950f29f37b" (UID: "4b05a676-12ac-429d-9e28-f5950f29f37b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.792383 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b05a676-12ac-429d-9e28-f5950f29f37b-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.792419 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e48ead74-9fba-4462-88f9-e1207f28d880-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:57 crc kubenswrapper[4715]: I1125 12:11:57.792429 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b05a676-12ac-429d-9e28-f5950f29f37b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.181706 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4b05a676-12ac-429d-9e28-f5950f29f37b","Type":"ContainerDied","Data":"0cd250a4823a894ac06f579c50323cea747a8b2e83804f1fec657d2ba0c2b0d7"} Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.182026 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cd250a4823a894ac06f579c50323cea747a8b2e83804f1fec657d2ba0c2b0d7" Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.181792 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.185206 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e48ead74-9fba-4462-88f9-e1207f28d880","Type":"ContainerDied","Data":"ed7cc62c7bd1ebe99011684e0386b6fc4644744a43b4c0e5251cbdbf41ee67d2"} Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.185243 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.185247 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed7cc62c7bd1ebe99011684e0386b6fc4644744a43b4c0e5251cbdbf41ee67d2" Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.591968 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.600070 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-dl4rf" Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.670799 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:58 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:58 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:58 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:58 crc kubenswrapper[4715]: I1125 12:11:58.670863 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:59 crc kubenswrapper[4715]: I1125 12:11:59.317904 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:59 crc kubenswrapper[4715]: I1125 12:11:59.322295 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3874a0dc-1a53-4587-bb14-7374a9715ae5-metrics-certs\") pod \"network-metrics-daemon-c8s6d\" (UID: \"3874a0dc-1a53-4587-bb14-7374a9715ae5\") " pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:59 crc kubenswrapper[4715]: I1125 12:11:59.373843 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-p6m6d" Nov 25 12:11:59 crc kubenswrapper[4715]: I1125 12:11:59.614524 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-c8s6d" Nov 25 12:11:59 crc kubenswrapper[4715]: I1125 12:11:59.664380 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:11:59 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:11:59 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:11:59 crc kubenswrapper[4715]: healthz check failed Nov 25 12:11:59 crc kubenswrapper[4715]: I1125 12:11:59.664467 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:11:59 crc kubenswrapper[4715]: I1125 12:11:59.834749 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-c8s6d"] Nov 25 12:11:59 crc kubenswrapper[4715]: W1125 12:11:59.868145 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3874a0dc_1a53_4587_bb14_7374a9715ae5.slice/crio-8029a1bae310456ff916cf6fdaa07007007d80dc7f60f4c21a57b41935332c25 WatchSource:0}: Error finding container 8029a1bae310456ff916cf6fdaa07007007d80dc7f60f4c21a57b41935332c25: Status 404 returned error can't find the container with id 8029a1bae310456ff916cf6fdaa07007007d80dc7f60f4c21a57b41935332c25 Nov 25 12:12:00 crc kubenswrapper[4715]: I1125 12:12:00.214653 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" event={"ID":"3874a0dc-1a53-4587-bb14-7374a9715ae5","Type":"ContainerStarted","Data":"8029a1bae310456ff916cf6fdaa07007007d80dc7f60f4c21a57b41935332c25"} Nov 25 12:12:00 crc kubenswrapper[4715]: I1125 12:12:00.664485 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:12:00 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:12:00 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:12:00 crc kubenswrapper[4715]: healthz check failed Nov 25 12:12:00 crc kubenswrapper[4715]: I1125 12:12:00.664580 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:12:01 crc kubenswrapper[4715]: I1125 12:12:01.227304 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" event={"ID":"3874a0dc-1a53-4587-bb14-7374a9715ae5","Type":"ContainerStarted","Data":"8c76dc007a6ee0af2f94497e9427a018861dbb4d21db50ed1883b3a6fe46e906"} Nov 25 12:12:01 crc kubenswrapper[4715]: I1125 12:12:01.662863 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:12:01 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:12:01 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:12:01 crc kubenswrapper[4715]: healthz check failed Nov 25 12:12:01 crc kubenswrapper[4715]: I1125 12:12:01.662919 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:12:02 crc kubenswrapper[4715]: I1125 12:12:02.663416 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:12:02 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:12:02 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:12:02 crc kubenswrapper[4715]: healthz check failed Nov 25 12:12:02 crc kubenswrapper[4715]: I1125 12:12:02.663771 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:12:03 crc kubenswrapper[4715]: I1125 12:12:03.662333 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:12:03 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:12:03 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:12:03 crc kubenswrapper[4715]: healthz check failed Nov 25 12:12:03 crc kubenswrapper[4715]: I1125 12:12:03.662389 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:12:03 crc kubenswrapper[4715]: I1125 12:12:03.832266 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-tvx5q container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 25 12:12:03 crc kubenswrapper[4715]: I1125 12:12:03.832331 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-tvx5q" podUID="13c92ba7-a34e-4f49-b98d-032d6d8543db" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 25 12:12:03 crc kubenswrapper[4715]: I1125 12:12:03.832348 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-tvx5q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 25 12:12:03 crc kubenswrapper[4715]: I1125 12:12:03.832405 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tvx5q" podUID="13c92ba7-a34e-4f49-b98d-032d6d8543db" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 25 12:12:03 crc kubenswrapper[4715]: I1125 12:12:03.841239 4715 patch_prober.go:28] interesting pod/console-f9d7485db-4dfdh container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Nov 25 12:12:03 crc kubenswrapper[4715]: I1125 12:12:03.841334 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-4dfdh" podUID="25b7b306-b4b9-4a46-bc4d-38df4f035c9f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Nov 25 12:12:04 crc kubenswrapper[4715]: I1125 12:12:04.663989 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:12:04 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Nov 25 12:12:04 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:12:04 crc kubenswrapper[4715]: healthz check failed Nov 25 12:12:04 crc kubenswrapper[4715]: I1125 12:12:04.664235 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:12:05 crc kubenswrapper[4715]: I1125 12:12:05.662856 4715 patch_prober.go:28] interesting pod/router-default-5444994796-vn6lh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 12:12:05 crc kubenswrapper[4715]: [+]has-synced ok Nov 25 12:12:05 crc kubenswrapper[4715]: [+]process-running ok Nov 25 12:12:05 crc kubenswrapper[4715]: healthz check failed Nov 25 12:12:05 crc kubenswrapper[4715]: I1125 12:12:05.662921 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vn6lh" podUID="2353d56f-1cab-4181-b1c2-164a90efab5e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 12:12:06 crc kubenswrapper[4715]: I1125 12:12:06.665952 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:12:06 crc kubenswrapper[4715]: I1125 12:12:06.669829 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-vn6lh" Nov 25 12:12:13 crc kubenswrapper[4715]: I1125 12:12:13.594084 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:12:13 crc kubenswrapper[4715]: I1125 12:12:13.844333 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-tvx5q" Nov 25 12:12:13 crc kubenswrapper[4715]: I1125 12:12:13.858579 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:12:13 crc kubenswrapper[4715]: I1125 12:12:13.867257 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-4dfdh" Nov 25 12:12:17 crc kubenswrapper[4715]: I1125 12:12:17.944363 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:12:17 crc kubenswrapper[4715]: I1125 12:12:17.944678 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:12:21 crc kubenswrapper[4715]: I1125 12:12:21.290302 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 12:12:24 crc kubenswrapper[4715]: I1125 12:12:24.350589 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-26dsn" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.683685 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 12:12:27 crc kubenswrapper[4715]: E1125 12:12:27.684590 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48ead74-9fba-4462-88f9-e1207f28d880" containerName="pruner" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.684616 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48ead74-9fba-4462-88f9-e1207f28d880" containerName="pruner" Nov 25 12:12:27 crc kubenswrapper[4715]: E1125 12:12:27.684683 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b05a676-12ac-429d-9e28-f5950f29f37b" containerName="pruner" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.684692 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b05a676-12ac-429d-9e28-f5950f29f37b" containerName="pruner" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.684826 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b05a676-12ac-429d-9e28-f5950f29f37b" containerName="pruner" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.684843 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e48ead74-9fba-4462-88f9-e1207f28d880" containerName="pruner" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.685881 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.688813 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.688858 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.699087 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.774586 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/328ca63d-6af9-40af-83a3-003f83f5ab6e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"328ca63d-6af9-40af-83a3-003f83f5ab6e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.774718 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/328ca63d-6af9-40af-83a3-003f83f5ab6e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"328ca63d-6af9-40af-83a3-003f83f5ab6e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.876233 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/328ca63d-6af9-40af-83a3-003f83f5ab6e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"328ca63d-6af9-40af-83a3-003f83f5ab6e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.876306 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/328ca63d-6af9-40af-83a3-003f83f5ab6e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"328ca63d-6af9-40af-83a3-003f83f5ab6e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.876397 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/328ca63d-6af9-40af-83a3-003f83f5ab6e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"328ca63d-6af9-40af-83a3-003f83f5ab6e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:27 crc kubenswrapper[4715]: I1125 12:12:27.900201 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/328ca63d-6af9-40af-83a3-003f83f5ab6e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"328ca63d-6af9-40af-83a3-003f83f5ab6e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:28 crc kubenswrapper[4715]: I1125 12:12:28.005466 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.278361 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.280094 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.295378 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.337702 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kube-api-access\") pod \"installer-9-crc\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.337775 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.337809 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-var-lock\") pod \"installer-9-crc\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.439377 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kube-api-access\") pod \"installer-9-crc\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.439712 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.439737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-var-lock\") pod \"installer-9-crc\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.439849 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-var-lock\") pod \"installer-9-crc\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.439889 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.457845 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kube-api-access\") pod \"installer-9-crc\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:32 crc kubenswrapper[4715]: E1125 12:12:32.577627 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 12:12:32 crc kubenswrapper[4715]: E1125 12:12:32.577798 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-btcqn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-wn6k5_openshift-marketplace(2b87d8f5-eda7-422d-ae7b-502725559317): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 12:12:32 crc kubenswrapper[4715]: E1125 12:12:32.579016 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-wn6k5" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" Nov 25 12:12:32 crc kubenswrapper[4715]: I1125 12:12:32.631447 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:12:43 crc kubenswrapper[4715]: E1125 12:12:43.228018 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 12:12:43 crc kubenswrapper[4715]: E1125 12:12:43.228867 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b4k5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rgjtn_openshift-marketplace(b3d2d042-e214-44ff-9c25-9d894bab99f3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 12:12:43 crc kubenswrapper[4715]: E1125 12:12:43.230997 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rgjtn" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" Nov 25 12:12:43 crc kubenswrapper[4715]: E1125 12:12:43.416051 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 12:12:43 crc kubenswrapper[4715]: E1125 12:12:43.416247 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r55px,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zxjdm_openshift-marketplace(a3a35017-fd36-43de-83bd-8787fece173a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 12:12:43 crc kubenswrapper[4715]: E1125 12:12:43.417623 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zxjdm" podUID="a3a35017-fd36-43de-83bd-8787fece173a" Nov 25 12:12:44 crc kubenswrapper[4715]: E1125 12:12:44.470616 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zxjdm" podUID="a3a35017-fd36-43de-83bd-8787fece173a" Nov 25 12:12:44 crc kubenswrapper[4715]: E1125 12:12:44.470703 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rgjtn" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" Nov 25 12:12:44 crc kubenswrapper[4715]: E1125 12:12:44.660176 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 12:12:44 crc kubenswrapper[4715]: E1125 12:12:44.660501 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jdbfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-z2nhl_openshift-marketplace(cc4608d5-28b1-4b1a-94da-72f4f7bb7907): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 12:12:44 crc kubenswrapper[4715]: E1125 12:12:44.661732 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-z2nhl" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" Nov 25 12:12:44 crc kubenswrapper[4715]: E1125 12:12:44.808402 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 12:12:44 crc kubenswrapper[4715]: E1125 12:12:44.810026 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6wx96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mwc6v_openshift-marketplace(55c59d7a-0cdb-48e7-a315-f2a3fdeee4df): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 12:12:44 crc kubenswrapper[4715]: E1125 12:12:44.811369 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mwc6v" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" Nov 25 12:12:45 crc kubenswrapper[4715]: I1125 12:12:45.022252 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 12:12:45 crc kubenswrapper[4715]: I1125 12:12:45.086792 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.144754 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.144970 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75wb8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2bpdr_openshift-marketplace(6da3d717-99fa-4202-909e-45db02ba3385): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.146708 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-2bpdr" podUID="6da3d717-99fa-4202-909e-45db02ba3385" Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.304002 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.304404 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vjv22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gdrl9_openshift-marketplace(6af7e329-2107-4548-b087-77b8b60f6646): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.305652 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gdrl9" podUID="6af7e329-2107-4548-b087-77b8b60f6646" Nov 25 12:12:45 crc kubenswrapper[4715]: I1125 12:12:45.484539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-c8s6d" event={"ID":"3874a0dc-1a53-4587-bb14-7374a9715ae5","Type":"ContainerStarted","Data":"57e630d3ee8d23083e6f3024c96556a74cb22bffbbc02cb845b97938e57a8ba1"} Nov 25 12:12:45 crc kubenswrapper[4715]: I1125 12:12:45.488453 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"328ca63d-6af9-40af-83a3-003f83f5ab6e","Type":"ContainerStarted","Data":"63c86ae0c4f249692f0916794fe51ac32b43f903b4a608ffc077bf6be9c4ee98"} Nov 25 12:12:45 crc kubenswrapper[4715]: I1125 12:12:45.492910 4715 generic.go:334] "Generic (PLEG): container finished" podID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerID="582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f" exitCode=0 Nov 25 12:12:45 crc kubenswrapper[4715]: I1125 12:12:45.493066 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l7rv8" event={"ID":"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3","Type":"ContainerDied","Data":"582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f"} Nov 25 12:12:45 crc kubenswrapper[4715]: I1125 12:12:45.495040 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2","Type":"ContainerStarted","Data":"0c4c299b22b50fb777591366e0065dc9f8748596899ac28bb759fea4b1f2d016"} Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.496386 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gdrl9" podUID="6af7e329-2107-4548-b087-77b8b60f6646" Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.497057 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mwc6v" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.497148 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-z2nhl" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" Nov 25 12:12:45 crc kubenswrapper[4715]: E1125 12:12:45.501278 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-2bpdr" podUID="6da3d717-99fa-4202-909e-45db02ba3385" Nov 25 12:12:45 crc kubenswrapper[4715]: I1125 12:12:45.510067 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-c8s6d" podStartSLOduration=189.510044018 podStartE2EDuration="3m9.510044018s" podCreationTimestamp="2025-11-25 12:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:12:45.501408615 +0000 UTC m=+216.008911646" watchObservedRunningTime="2025-11-25 12:12:45.510044018 +0000 UTC m=+216.017547039" Nov 25 12:12:46 crc kubenswrapper[4715]: I1125 12:12:46.505618 4715 generic.go:334] "Generic (PLEG): container finished" podID="2b87d8f5-eda7-422d-ae7b-502725559317" containerID="46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184" exitCode=0 Nov 25 12:12:46 crc kubenswrapper[4715]: I1125 12:12:46.505769 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wn6k5" event={"ID":"2b87d8f5-eda7-422d-ae7b-502725559317","Type":"ContainerDied","Data":"46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184"} Nov 25 12:12:46 crc kubenswrapper[4715]: I1125 12:12:46.508649 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2","Type":"ContainerStarted","Data":"52fd2fe0a4942b9ac1d15de7547dc8b0d30b1c8f6fb2a82f8c0659d539f10576"} Nov 25 12:12:46 crc kubenswrapper[4715]: I1125 12:12:46.512568 4715 generic.go:334] "Generic (PLEG): container finished" podID="328ca63d-6af9-40af-83a3-003f83f5ab6e" containerID="0570956920d0ebeae2fd9b0e813ccec83c41d5cc0d423e09ffa9bcff9e2959c4" exitCode=0 Nov 25 12:12:46 crc kubenswrapper[4715]: I1125 12:12:46.513802 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"328ca63d-6af9-40af-83a3-003f83f5ab6e","Type":"ContainerDied","Data":"0570956920d0ebeae2fd9b0e813ccec83c41d5cc0d423e09ffa9bcff9e2959c4"} Nov 25 12:12:46 crc kubenswrapper[4715]: I1125 12:12:46.546987 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=14.54693644 podStartE2EDuration="14.54693644s" podCreationTimestamp="2025-11-25 12:12:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:12:46.546238909 +0000 UTC m=+217.053741950" watchObservedRunningTime="2025-11-25 12:12:46.54693644 +0000 UTC m=+217.054439461" Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.524536 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l7rv8" event={"ID":"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3","Type":"ContainerStarted","Data":"292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4"} Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.545693 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l7rv8" podStartSLOduration=3.452081573 podStartE2EDuration="57.545672138s" podCreationTimestamp="2025-11-25 12:11:50 +0000 UTC" firstStartedPulling="2025-11-25 12:11:52.992575764 +0000 UTC m=+163.500078785" lastFinishedPulling="2025-11-25 12:12:47.086166319 +0000 UTC m=+217.593669350" observedRunningTime="2025-11-25 12:12:47.541210343 +0000 UTC m=+218.048713374" watchObservedRunningTime="2025-11-25 12:12:47.545672138 +0000 UTC m=+218.053175159" Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.757778 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.883782 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/328ca63d-6af9-40af-83a3-003f83f5ab6e-kube-api-access\") pod \"328ca63d-6af9-40af-83a3-003f83f5ab6e\" (UID: \"328ca63d-6af9-40af-83a3-003f83f5ab6e\") " Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.883820 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/328ca63d-6af9-40af-83a3-003f83f5ab6e-kubelet-dir\") pod \"328ca63d-6af9-40af-83a3-003f83f5ab6e\" (UID: \"328ca63d-6af9-40af-83a3-003f83f5ab6e\") " Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.884046 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/328ca63d-6af9-40af-83a3-003f83f5ab6e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "328ca63d-6af9-40af-83a3-003f83f5ab6e" (UID: "328ca63d-6af9-40af-83a3-003f83f5ab6e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.889065 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/328ca63d-6af9-40af-83a3-003f83f5ab6e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "328ca63d-6af9-40af-83a3-003f83f5ab6e" (UID: "328ca63d-6af9-40af-83a3-003f83f5ab6e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.944151 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.944231 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.944294 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.944824 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71"} pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.944981 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" containerID="cri-o://2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71" gracePeriod=600 Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.984861 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/328ca63d-6af9-40af-83a3-003f83f5ab6e-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 12:12:47 crc kubenswrapper[4715]: I1125 12:12:47.984895 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/328ca63d-6af9-40af-83a3-003f83f5ab6e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:12:48 crc kubenswrapper[4715]: I1125 12:12:48.532018 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"328ca63d-6af9-40af-83a3-003f83f5ab6e","Type":"ContainerDied","Data":"63c86ae0c4f249692f0916794fe51ac32b43f903b4a608ffc077bf6be9c4ee98"} Nov 25 12:12:48 crc kubenswrapper[4715]: I1125 12:12:48.532342 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63c86ae0c4f249692f0916794fe51ac32b43f903b4a608ffc077bf6be9c4ee98" Nov 25 12:12:48 crc kubenswrapper[4715]: I1125 12:12:48.532075 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 12:12:48 crc kubenswrapper[4715]: I1125 12:12:48.534031 4715 generic.go:334] "Generic (PLEG): container finished" podID="51676837-adc4-4424-b527-920a6528b6a2" containerID="2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71" exitCode=0 Nov 25 12:12:48 crc kubenswrapper[4715]: I1125 12:12:48.534103 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerDied","Data":"2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71"} Nov 25 12:12:48 crc kubenswrapper[4715]: I1125 12:12:48.536481 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wn6k5" event={"ID":"2b87d8f5-eda7-422d-ae7b-502725559317","Type":"ContainerStarted","Data":"beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b"} Nov 25 12:12:49 crc kubenswrapper[4715]: I1125 12:12:49.542890 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"133f6a3b52123e9fe263e667ae6646c6fa53439cce70b559acdf67b4d86c9ee6"} Nov 25 12:12:49 crc kubenswrapper[4715]: I1125 12:12:49.561802 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wn6k5" podStartSLOduration=3.600462683 podStartE2EDuration="56.56178308s" podCreationTimestamp="2025-11-25 12:11:53 +0000 UTC" firstStartedPulling="2025-11-25 12:11:55.135525357 +0000 UTC m=+165.643028378" lastFinishedPulling="2025-11-25 12:12:48.096845754 +0000 UTC m=+218.604348775" observedRunningTime="2025-11-25 12:12:48.560464418 +0000 UTC m=+219.067967439" watchObservedRunningTime="2025-11-25 12:12:49.56178308 +0000 UTC m=+220.069286121" Nov 25 12:12:51 crc kubenswrapper[4715]: I1125 12:12:51.385903 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:12:51 crc kubenswrapper[4715]: I1125 12:12:51.386547 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:12:51 crc kubenswrapper[4715]: I1125 12:12:51.523253 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:12:54 crc kubenswrapper[4715]: I1125 12:12:54.076114 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:12:54 crc kubenswrapper[4715]: I1125 12:12:54.076686 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:12:54 crc kubenswrapper[4715]: I1125 12:12:54.120604 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:12:54 crc kubenswrapper[4715]: I1125 12:12:54.603979 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:12:55 crc kubenswrapper[4715]: I1125 12:12:55.735634 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wn6k5"] Nov 25 12:12:56 crc kubenswrapper[4715]: I1125 12:12:56.577133 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wn6k5" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" containerName="registry-server" containerID="cri-o://beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b" gracePeriod=2 Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.093645 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.232978 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-utilities\") pod \"2b87d8f5-eda7-422d-ae7b-502725559317\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.233062 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btcqn\" (UniqueName: \"kubernetes.io/projected/2b87d8f5-eda7-422d-ae7b-502725559317-kube-api-access-btcqn\") pod \"2b87d8f5-eda7-422d-ae7b-502725559317\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.233233 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-catalog-content\") pod \"2b87d8f5-eda7-422d-ae7b-502725559317\" (UID: \"2b87d8f5-eda7-422d-ae7b-502725559317\") " Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.235289 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-utilities" (OuterVolumeSpecName: "utilities") pod "2b87d8f5-eda7-422d-ae7b-502725559317" (UID: "2b87d8f5-eda7-422d-ae7b-502725559317"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.238924 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b87d8f5-eda7-422d-ae7b-502725559317-kube-api-access-btcqn" (OuterVolumeSpecName: "kube-api-access-btcqn") pod "2b87d8f5-eda7-422d-ae7b-502725559317" (UID: "2b87d8f5-eda7-422d-ae7b-502725559317"). InnerVolumeSpecName "kube-api-access-btcqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.335132 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.335690 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btcqn\" (UniqueName: \"kubernetes.io/projected/2b87d8f5-eda7-422d-ae7b-502725559317-kube-api-access-btcqn\") on node \"crc\" DevicePath \"\"" Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.595585 4715 generic.go:334] "Generic (PLEG): container finished" podID="2b87d8f5-eda7-422d-ae7b-502725559317" containerID="beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b" exitCode=0 Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.595629 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wn6k5" event={"ID":"2b87d8f5-eda7-422d-ae7b-502725559317","Type":"ContainerDied","Data":"beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b"} Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.595650 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wn6k5" Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.595665 4715 scope.go:117] "RemoveContainer" containerID="beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b" Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.595655 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wn6k5" event={"ID":"2b87d8f5-eda7-422d-ae7b-502725559317","Type":"ContainerDied","Data":"baea65f3b3ca21551daf20ddd333918d5e445e57dc763ee1117b907903e9aab0"} Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.909248 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b87d8f5-eda7-422d-ae7b-502725559317" (UID: "2b87d8f5-eda7-422d-ae7b-502725559317"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:12:59 crc kubenswrapper[4715]: I1125 12:12:59.943718 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b87d8f5-eda7-422d-ae7b-502725559317-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.088395 4715 scope.go:117] "RemoveContainer" containerID="46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184" Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.131898 4715 scope.go:117] "RemoveContainer" containerID="31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263" Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.193547 4715 scope.go:117] "RemoveContainer" containerID="beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b" Nov 25 12:13:00 crc kubenswrapper[4715]: E1125 12:13:00.194730 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b\": container with ID starting with beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b not found: ID does not exist" containerID="beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b" Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.194763 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b"} err="failed to get container status \"beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b\": rpc error: code = NotFound desc = could not find container \"beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b\": container with ID starting with beb56c216112fbf912f061d8ebdcd1c6d77e12accce7c232513d91fd7d375f8b not found: ID does not exist" Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.194785 4715 scope.go:117] "RemoveContainer" containerID="46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184" Nov 25 12:13:00 crc kubenswrapper[4715]: E1125 12:13:00.195300 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184\": container with ID starting with 46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184 not found: ID does not exist" containerID="46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184" Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.195333 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184"} err="failed to get container status \"46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184\": rpc error: code = NotFound desc = could not find container \"46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184\": container with ID starting with 46555b4785345bedcb7586567e57028cba66be3b1ad11e04cfd7aee35686c184 not found: ID does not exist" Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.195353 4715 scope.go:117] "RemoveContainer" containerID="31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263" Nov 25 12:13:00 crc kubenswrapper[4715]: E1125 12:13:00.196034 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263\": container with ID starting with 31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263 not found: ID does not exist" containerID="31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263" Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.196065 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263"} err="failed to get container status \"31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263\": rpc error: code = NotFound desc = could not find container \"31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263\": container with ID starting with 31ef2d62c706224261df773c4f3f4f88f73ede45093935c22137b263d84ac263 not found: ID does not exist" Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.228480 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wn6k5"] Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.232669 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wn6k5"] Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.606725 4715 generic.go:334] "Generic (PLEG): container finished" podID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerID="cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31" exitCode=0 Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.607033 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwc6v" event={"ID":"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df","Type":"ContainerDied","Data":"cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31"} Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.610252 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2nhl" event={"ID":"cc4608d5-28b1-4b1a-94da-72f4f7bb7907","Type":"ContainerStarted","Data":"fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d"} Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.612407 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rgjtn" event={"ID":"b3d2d042-e214-44ff-9c25-9d894bab99f3","Type":"ContainerStarted","Data":"1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9"} Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.614286 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bpdr" event={"ID":"6da3d717-99fa-4202-909e-45db02ba3385","Type":"ContainerStarted","Data":"10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4"} Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.616087 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdrl9" event={"ID":"6af7e329-2107-4548-b087-77b8b60f6646","Type":"ContainerStarted","Data":"1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43"} Nov 25 12:13:00 crc kubenswrapper[4715]: I1125 12:13:00.700897 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" path="/var/lib/kubelet/pods/2b87d8f5-eda7-422d-ae7b-502725559317/volumes" Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.427428 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.625010 4715 generic.go:334] "Generic (PLEG): container finished" podID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerID="fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d" exitCode=0 Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.625086 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2nhl" event={"ID":"cc4608d5-28b1-4b1a-94da-72f4f7bb7907","Type":"ContainerDied","Data":"fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d"} Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.629008 4715 generic.go:334] "Generic (PLEG): container finished" podID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerID="1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9" exitCode=0 Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.629069 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rgjtn" event={"ID":"b3d2d042-e214-44ff-9c25-9d894bab99f3","Type":"ContainerDied","Data":"1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9"} Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.630722 4715 generic.go:334] "Generic (PLEG): container finished" podID="6da3d717-99fa-4202-909e-45db02ba3385" containerID="10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4" exitCode=0 Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.630784 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bpdr" event={"ID":"6da3d717-99fa-4202-909e-45db02ba3385","Type":"ContainerDied","Data":"10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4"} Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.632635 4715 generic.go:334] "Generic (PLEG): container finished" podID="6af7e329-2107-4548-b087-77b8b60f6646" containerID="1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43" exitCode=0 Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.632656 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdrl9" event={"ID":"6af7e329-2107-4548-b087-77b8b60f6646","Type":"ContainerDied","Data":"1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43"} Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.634871 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwc6v" event={"ID":"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df","Type":"ContainerStarted","Data":"87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65"} Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.640302 4715 generic.go:334] "Generic (PLEG): container finished" podID="a3a35017-fd36-43de-83bd-8787fece173a" containerID="32759f7febe258537e90b6193ea71ca32d0566c5dd5dd7428d6cd87ea73552ef" exitCode=0 Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.640370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxjdm" event={"ID":"a3a35017-fd36-43de-83bd-8787fece173a","Type":"ContainerDied","Data":"32759f7febe258537e90b6193ea71ca32d0566c5dd5dd7428d6cd87ea73552ef"} Nov 25 12:13:01 crc kubenswrapper[4715]: I1125 12:13:01.677211 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mwc6v" podStartSLOduration=2.635461513 podStartE2EDuration="1m9.677150715s" podCreationTimestamp="2025-11-25 12:11:52 +0000 UTC" firstStartedPulling="2025-11-25 12:11:54.053387518 +0000 UTC m=+164.560890539" lastFinishedPulling="2025-11-25 12:13:01.09507672 +0000 UTC m=+231.602579741" observedRunningTime="2025-11-25 12:13:01.67415088 +0000 UTC m=+232.181653921" watchObservedRunningTime="2025-11-25 12:13:01.677150715 +0000 UTC m=+232.184653736" Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.533501 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l7rv8"] Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.533751 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l7rv8" podUID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerName="registry-server" containerID="cri-o://292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4" gracePeriod=2 Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.647958 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2nhl" event={"ID":"cc4608d5-28b1-4b1a-94da-72f4f7bb7907","Type":"ContainerStarted","Data":"5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844"} Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.652582 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxjdm" event={"ID":"a3a35017-fd36-43de-83bd-8787fece173a","Type":"ContainerStarted","Data":"97d2f720f03fff20718215ecdc9f08bbac2ef1bf54d79efc33731ba620e1b706"} Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.658212 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rgjtn" event={"ID":"b3d2d042-e214-44ff-9c25-9d894bab99f3","Type":"ContainerStarted","Data":"7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17"} Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.659923 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bpdr" event={"ID":"6da3d717-99fa-4202-909e-45db02ba3385","Type":"ContainerStarted","Data":"fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56"} Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.669680 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z2nhl" podStartSLOduration=2.349835728 podStartE2EDuration="1m12.669662568s" podCreationTimestamp="2025-11-25 12:11:50 +0000 UTC" firstStartedPulling="2025-11-25 12:11:51.910023253 +0000 UTC m=+162.417526274" lastFinishedPulling="2025-11-25 12:13:02.229850093 +0000 UTC m=+232.737353114" observedRunningTime="2025-11-25 12:13:02.669260077 +0000 UTC m=+233.176763098" watchObservedRunningTime="2025-11-25 12:13:02.669662568 +0000 UTC m=+233.177165589" Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.685583 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdrl9" event={"ID":"6af7e329-2107-4548-b087-77b8b60f6646","Type":"ContainerStarted","Data":"c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519"} Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.693151 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2bpdr" podStartSLOduration=2.739789213 podStartE2EDuration="1m9.693132958s" podCreationTimestamp="2025-11-25 12:11:53 +0000 UTC" firstStartedPulling="2025-11-25 12:11:55.122607911 +0000 UTC m=+165.630110942" lastFinishedPulling="2025-11-25 12:13:02.075951666 +0000 UTC m=+232.583454687" observedRunningTime="2025-11-25 12:13:02.691498612 +0000 UTC m=+233.199001653" watchObservedRunningTime="2025-11-25 12:13:02.693132958 +0000 UTC m=+233.200635979" Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.716899 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zxjdm" podStartSLOduration=3.571832851 podStartE2EDuration="1m12.716874186s" podCreationTimestamp="2025-11-25 12:11:50 +0000 UTC" firstStartedPulling="2025-11-25 12:11:53.002740431 +0000 UTC m=+163.510243452" lastFinishedPulling="2025-11-25 12:13:02.147781766 +0000 UTC m=+232.655284787" observedRunningTime="2025-11-25 12:13:02.712442451 +0000 UTC m=+233.219945482" watchObservedRunningTime="2025-11-25 12:13:02.716874186 +0000 UTC m=+233.224377207" Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.730959 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rgjtn" podStartSLOduration=1.378420016 podStartE2EDuration="1m10.730942591s" podCreationTimestamp="2025-11-25 12:11:52 +0000 UTC" firstStartedPulling="2025-11-25 12:11:52.962624866 +0000 UTC m=+163.470127897" lastFinishedPulling="2025-11-25 12:13:02.315147451 +0000 UTC m=+232.822650472" observedRunningTime="2025-11-25 12:13:02.728632706 +0000 UTC m=+233.236135727" watchObservedRunningTime="2025-11-25 12:13:02.730942591 +0000 UTC m=+233.238445612" Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.744968 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gdrl9" podStartSLOduration=2.463230186 podStartE2EDuration="1m12.744949895s" podCreationTimestamp="2025-11-25 12:11:50 +0000 UTC" firstStartedPulling="2025-11-25 12:11:51.907030098 +0000 UTC m=+162.414533119" lastFinishedPulling="2025-11-25 12:13:02.188749807 +0000 UTC m=+232.696252828" observedRunningTime="2025-11-25 12:13:02.744590565 +0000 UTC m=+233.252093586" watchObservedRunningTime="2025-11-25 12:13:02.744949895 +0000 UTC m=+233.252452916" Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.850003 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.850757 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.933017 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:13:02 crc kubenswrapper[4715]: I1125 12:13:02.946389 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.091679 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-catalog-content\") pod \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.091728 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5w78\" (UniqueName: \"kubernetes.io/projected/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-kube-api-access-l5w78\") pod \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.091756 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-utilities\") pod \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\" (UID: \"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3\") " Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.092591 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-utilities" (OuterVolumeSpecName: "utilities") pod "ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" (UID: "ce959404-ce4f-4c31-8d7b-4fd0c4b458f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.097379 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-kube-api-access-l5w78" (OuterVolumeSpecName: "kube-api-access-l5w78") pod "ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" (UID: "ce959404-ce4f-4c31-8d7b-4fd0c4b458f3"). InnerVolumeSpecName "kube-api-access-l5w78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.151534 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" (UID: "ce959404-ce4f-4c31-8d7b-4fd0c4b458f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.193263 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.193318 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5w78\" (UniqueName: \"kubernetes.io/projected/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-kube-api-access-l5w78\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.193331 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.674889 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.675015 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.694889 4715 generic.go:334] "Generic (PLEG): container finished" podID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerID="292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4" exitCode=0 Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.696538 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l7rv8" event={"ID":"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3","Type":"ContainerDied","Data":"292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4"} Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.696601 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l7rv8" event={"ID":"ce959404-ce4f-4c31-8d7b-4fd0c4b458f3","Type":"ContainerDied","Data":"d75834ab177f8e867abc49607b9dd8c591e1b452b004f5a3753647c2faba3dcc"} Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.696628 4715 scope.go:117] "RemoveContainer" containerID="292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.696862 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l7rv8" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.720396 4715 scope.go:117] "RemoveContainer" containerID="582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.750061 4715 scope.go:117] "RemoveContainer" containerID="b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.769853 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l7rv8"] Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.772266 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l7rv8"] Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.789401 4715 scope.go:117] "RemoveContainer" containerID="292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4" Nov 25 12:13:03 crc kubenswrapper[4715]: E1125 12:13:03.792272 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4\": container with ID starting with 292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4 not found: ID does not exist" containerID="292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.792329 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4"} err="failed to get container status \"292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4\": rpc error: code = NotFound desc = could not find container \"292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4\": container with ID starting with 292283ba9152cc7570afad9797ff35eaf232afe4ccd7aeea66603d516d34bff4 not found: ID does not exist" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.792367 4715 scope.go:117] "RemoveContainer" containerID="582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f" Nov 25 12:13:03 crc kubenswrapper[4715]: E1125 12:13:03.792975 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f\": container with ID starting with 582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f not found: ID does not exist" containerID="582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.793003 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f"} err="failed to get container status \"582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f\": rpc error: code = NotFound desc = could not find container \"582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f\": container with ID starting with 582e5fda748e0e59337d274c0043340228a9f365e5390b955bf434310e090e0f not found: ID does not exist" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.793021 4715 scope.go:117] "RemoveContainer" containerID="b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d" Nov 25 12:13:03 crc kubenswrapper[4715]: E1125 12:13:03.793366 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d\": container with ID starting with b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d not found: ID does not exist" containerID="b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d" Nov 25 12:13:03 crc kubenswrapper[4715]: I1125 12:13:03.793394 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d"} err="failed to get container status \"b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d\": rpc error: code = NotFound desc = could not find container \"b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d\": container with ID starting with b50b6b1cb844056118dcca23afad4c10f445ad0366f95da939a03343a179d12d not found: ID does not exist" Nov 25 12:13:04 crc kubenswrapper[4715]: I1125 12:13:04.700586 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" path="/var/lib/kubelet/pods/ce959404-ce4f-4c31-8d7b-4fd0c4b458f3/volumes" Nov 25 12:13:04 crc kubenswrapper[4715]: I1125 12:13:04.720168 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2bpdr" podUID="6da3d717-99fa-4202-909e-45db02ba3385" containerName="registry-server" probeResult="failure" output=< Nov 25 12:13:04 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Nov 25 12:13:04 crc kubenswrapper[4715]: > Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028126 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pc2s5"] Nov 25 12:13:08 crc kubenswrapper[4715]: E1125 12:13:08.028652 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" containerName="registry-server" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028664 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" containerName="registry-server" Nov 25 12:13:08 crc kubenswrapper[4715]: E1125 12:13:08.028673 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerName="registry-server" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028682 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerName="registry-server" Nov 25 12:13:08 crc kubenswrapper[4715]: E1125 12:13:08.028693 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerName="extract-content" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028701 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerName="extract-content" Nov 25 12:13:08 crc kubenswrapper[4715]: E1125 12:13:08.028719 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerName="extract-utilities" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028726 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerName="extract-utilities" Nov 25 12:13:08 crc kubenswrapper[4715]: E1125 12:13:08.028746 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" containerName="extract-content" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028752 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" containerName="extract-content" Nov 25 12:13:08 crc kubenswrapper[4715]: E1125 12:13:08.028763 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" containerName="extract-utilities" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028768 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" containerName="extract-utilities" Nov 25 12:13:08 crc kubenswrapper[4715]: E1125 12:13:08.028776 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="328ca63d-6af9-40af-83a3-003f83f5ab6e" containerName="pruner" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028783 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="328ca63d-6af9-40af-83a3-003f83f5ab6e" containerName="pruner" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028869 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="328ca63d-6af9-40af-83a3-003f83f5ab6e" containerName="pruner" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028879 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b87d8f5-eda7-422d-ae7b-502725559317" containerName="registry-server" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.028890 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce959404-ce4f-4c31-8d7b-4fd0c4b458f3" containerName="registry-server" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.029291 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.045353 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pc2s5"] Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.154785 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9j9v\" (UniqueName: \"kubernetes.io/projected/df09a836-93f2-4f75-96ab-41b2e0426785-kube-api-access-n9j9v\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.155101 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/df09a836-93f2-4f75-96ab-41b2e0426785-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.155125 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df09a836-93f2-4f75-96ab-41b2e0426785-bound-sa-token\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.155146 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df09a836-93f2-4f75-96ab-41b2e0426785-trusted-ca\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.155200 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.155223 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/df09a836-93f2-4f75-96ab-41b2e0426785-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.155243 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/df09a836-93f2-4f75-96ab-41b2e0426785-registry-certificates\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.155273 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/df09a836-93f2-4f75-96ab-41b2e0426785-registry-tls\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.192319 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.256804 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/df09a836-93f2-4f75-96ab-41b2e0426785-registry-certificates\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.256894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/df09a836-93f2-4f75-96ab-41b2e0426785-registry-tls\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.256933 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9j9v\" (UniqueName: \"kubernetes.io/projected/df09a836-93f2-4f75-96ab-41b2e0426785-kube-api-access-n9j9v\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.256976 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/df09a836-93f2-4f75-96ab-41b2e0426785-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.257005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df09a836-93f2-4f75-96ab-41b2e0426785-bound-sa-token\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.257059 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df09a836-93f2-4f75-96ab-41b2e0426785-trusted-ca\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.257841 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/df09a836-93f2-4f75-96ab-41b2e0426785-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.257400 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/df09a836-93f2-4f75-96ab-41b2e0426785-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.258503 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df09a836-93f2-4f75-96ab-41b2e0426785-trusted-ca\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.259082 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/df09a836-93f2-4f75-96ab-41b2e0426785-registry-certificates\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.273172 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/df09a836-93f2-4f75-96ab-41b2e0426785-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.273240 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/df09a836-93f2-4f75-96ab-41b2e0426785-registry-tls\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.275400 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9j9v\" (UniqueName: \"kubernetes.io/projected/df09a836-93f2-4f75-96ab-41b2e0426785-kube-api-access-n9j9v\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.278706 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df09a836-93f2-4f75-96ab-41b2e0426785-bound-sa-token\") pod \"image-registry-66df7c8f76-pc2s5\" (UID: \"df09a836-93f2-4f75-96ab-41b2e0426785\") " pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.349167 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:08 crc kubenswrapper[4715]: I1125 12:13:08.731790 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pc2s5"] Nov 25 12:13:08 crc kubenswrapper[4715]: W1125 12:13:08.741760 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf09a836_93f2_4f75_96ab_41b2e0426785.slice/crio-99663a8c989300b0f7283fe365af12bd1e52bde98d404196e3b294b43dec49d6 WatchSource:0}: Error finding container 99663a8c989300b0f7283fe365af12bd1e52bde98d404196e3b294b43dec49d6: Status 404 returned error can't find the container with id 99663a8c989300b0f7283fe365af12bd1e52bde98d404196e3b294b43dec49d6 Nov 25 12:13:09 crc kubenswrapper[4715]: I1125 12:13:09.735049 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" event={"ID":"df09a836-93f2-4f75-96ab-41b2e0426785","Type":"ContainerStarted","Data":"3a4fb5d898b300c388580f90ca3db96f254af45c85176cb87ff2a5c9e3d42585"} Nov 25 12:13:09 crc kubenswrapper[4715]: I1125 12:13:09.735448 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" event={"ID":"df09a836-93f2-4f75-96ab-41b2e0426785","Type":"ContainerStarted","Data":"99663a8c989300b0f7283fe365af12bd1e52bde98d404196e3b294b43dec49d6"} Nov 25 12:13:09 crc kubenswrapper[4715]: I1125 12:13:09.736367 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:09 crc kubenswrapper[4715]: I1125 12:13:09.756687 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" podStartSLOduration=1.756669514 podStartE2EDuration="1.756669514s" podCreationTimestamp="2025-11-25 12:13:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:13:09.754942596 +0000 UTC m=+240.262445617" watchObservedRunningTime="2025-11-25 12:13:09.756669514 +0000 UTC m=+240.264172535" Nov 25 12:13:10 crc kubenswrapper[4715]: I1125 12:13:10.677729 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:13:10 crc kubenswrapper[4715]: I1125 12:13:10.678033 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:13:10 crc kubenswrapper[4715]: I1125 12:13:10.734092 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:13:10 crc kubenswrapper[4715]: I1125 12:13:10.779082 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:13:11 crc kubenswrapper[4715]: I1125 12:13:11.085575 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:13:11 crc kubenswrapper[4715]: I1125 12:13:11.085621 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:13:11 crc kubenswrapper[4715]: I1125 12:13:11.130676 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:13:11 crc kubenswrapper[4715]: I1125 12:13:11.381266 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:13:11 crc kubenswrapper[4715]: I1125 12:13:11.381321 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:13:11 crc kubenswrapper[4715]: I1125 12:13:11.419306 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:13:11 crc kubenswrapper[4715]: I1125 12:13:11.789029 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:13:11 crc kubenswrapper[4715]: I1125 12:13:11.789100 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:13:12 crc kubenswrapper[4715]: I1125 12:13:12.364775 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2nhl"] Nov 25 12:13:12 crc kubenswrapper[4715]: I1125 12:13:12.452129 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:13:12 crc kubenswrapper[4715]: I1125 12:13:12.452201 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:13:12 crc kubenswrapper[4715]: I1125 12:13:12.490199 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:13:12 crc kubenswrapper[4715]: I1125 12:13:12.784088 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:13:12 crc kubenswrapper[4715]: I1125 12:13:12.892608 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.629713 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gdrl9"] Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.630825 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gdrl9" podUID="6af7e329-2107-4548-b087-77b8b60f6646" containerName="registry-server" containerID="cri-o://c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519" gracePeriod=30 Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.646351 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zxjdm"] Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.649906 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rstj"] Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.650529 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" podUID="6467e871-dd98-4bcf-bf7c-53726589346b" containerName="marketplace-operator" containerID="cri-o://edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303" gracePeriod=30 Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.662216 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwc6v"] Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.664899 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-25kht"] Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.665556 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.671769 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rgjtn"] Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.674669 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2bpdr"] Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.674909 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2bpdr" podUID="6da3d717-99fa-4202-909e-45db02ba3385" containerName="registry-server" containerID="cri-o://fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56" gracePeriod=30 Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.678464 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-25kht"] Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.742950 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c6da464-9fba-4591-a5b1-dfd37dde0770-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-25kht\" (UID: \"5c6da464-9fba-4591-a5b1-dfd37dde0770\") " pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.743032 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x92q6\" (UniqueName: \"kubernetes.io/projected/5c6da464-9fba-4591-a5b1-dfd37dde0770-kube-api-access-x92q6\") pod \"marketplace-operator-79b997595-25kht\" (UID: \"5c6da464-9fba-4591-a5b1-dfd37dde0770\") " pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.743096 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c6da464-9fba-4591-a5b1-dfd37dde0770-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-25kht\" (UID: \"5c6da464-9fba-4591-a5b1-dfd37dde0770\") " pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.751972 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zxjdm" podUID="a3a35017-fd36-43de-83bd-8787fece173a" containerName="registry-server" containerID="cri-o://97d2f720f03fff20718215ecdc9f08bbac2ef1bf54d79efc33731ba620e1b706" gracePeriod=30 Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.752396 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z2nhl" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerName="registry-server" containerID="cri-o://5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844" gracePeriod=2 Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.752644 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mwc6v" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerName="registry-server" containerID="cri-o://87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65" gracePeriod=30 Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.843929 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c6da464-9fba-4591-a5b1-dfd37dde0770-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-25kht\" (UID: \"5c6da464-9fba-4591-a5b1-dfd37dde0770\") " pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.844008 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x92q6\" (UniqueName: \"kubernetes.io/projected/5c6da464-9fba-4591-a5b1-dfd37dde0770-kube-api-access-x92q6\") pod \"marketplace-operator-79b997595-25kht\" (UID: \"5c6da464-9fba-4591-a5b1-dfd37dde0770\") " pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.844050 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c6da464-9fba-4591-a5b1-dfd37dde0770-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-25kht\" (UID: \"5c6da464-9fba-4591-a5b1-dfd37dde0770\") " pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.845462 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5c6da464-9fba-4591-a5b1-dfd37dde0770-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-25kht\" (UID: \"5c6da464-9fba-4591-a5b1-dfd37dde0770\") " pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.850367 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5c6da464-9fba-4591-a5b1-dfd37dde0770-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-25kht\" (UID: \"5c6da464-9fba-4591-a5b1-dfd37dde0770\") " pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.861241 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x92q6\" (UniqueName: \"kubernetes.io/projected/5c6da464-9fba-4591-a5b1-dfd37dde0770-kube-api-access-x92q6\") pod \"marketplace-operator-79b997595-25kht\" (UID: \"5c6da464-9fba-4591-a5b1-dfd37dde0770\") " pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.908027 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2rstj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.908085 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" podUID="6467e871-dd98-4bcf-bf7c-53726589346b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Nov 25 12:13:13 crc kubenswrapper[4715]: I1125 12:13:13.983980 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.410860 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.445304 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-25kht"] Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.448395 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.457366 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.464859 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wx96\" (UniqueName: \"kubernetes.io/projected/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-kube-api-access-6wx96\") pod \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.464922 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-utilities\") pod \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.464993 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-catalog-content\") pod \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\" (UID: \"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.467006 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-utilities" (OuterVolumeSpecName: "utilities") pod "55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" (UID: "55c59d7a-0cdb-48e7-a315-f2a3fdeee4df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.475949 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-kube-api-access-6wx96" (OuterVolumeSpecName: "kube-api-access-6wx96") pod "55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" (UID: "55c59d7a-0cdb-48e7-a315-f2a3fdeee4df"). InnerVolumeSpecName "kube-api-access-6wx96". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.568544 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-catalog-content\") pod \"6da3d717-99fa-4202-909e-45db02ba3385\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.568614 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdbfx\" (UniqueName: \"kubernetes.io/projected/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-kube-api-access-jdbfx\") pod \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.568671 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-utilities\") pod \"6da3d717-99fa-4202-909e-45db02ba3385\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.568697 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-utilities\") pod \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.568726 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-catalog-content\") pod \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\" (UID: \"cc4608d5-28b1-4b1a-94da-72f4f7bb7907\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.568753 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75wb8\" (UniqueName: \"kubernetes.io/projected/6da3d717-99fa-4202-909e-45db02ba3385-kube-api-access-75wb8\") pod \"6da3d717-99fa-4202-909e-45db02ba3385\" (UID: \"6da3d717-99fa-4202-909e-45db02ba3385\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.569085 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.569104 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wx96\" (UniqueName: \"kubernetes.io/projected/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-kube-api-access-6wx96\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.571754 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-utilities" (OuterVolumeSpecName: "utilities") pod "6da3d717-99fa-4202-909e-45db02ba3385" (UID: "6da3d717-99fa-4202-909e-45db02ba3385"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.574054 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-utilities" (OuterVolumeSpecName: "utilities") pod "cc4608d5-28b1-4b1a-94da-72f4f7bb7907" (UID: "cc4608d5-28b1-4b1a-94da-72f4f7bb7907"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.574413 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qztrg"] Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.579362 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da3d717-99fa-4202-909e-45db02ba3385-kube-api-access-75wb8" (OuterVolumeSpecName: "kube-api-access-75wb8") pod "6da3d717-99fa-4202-909e-45db02ba3385" (UID: "6da3d717-99fa-4202-909e-45db02ba3385"). InnerVolumeSpecName "kube-api-access-75wb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.582090 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-kube-api-access-jdbfx" (OuterVolumeSpecName: "kube-api-access-jdbfx") pod "cc4608d5-28b1-4b1a-94da-72f4f7bb7907" (UID: "cc4608d5-28b1-4b1a-94da-72f4f7bb7907"). InnerVolumeSpecName "kube-api-access-jdbfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.647056 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.669231 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" (UID: "55c59d7a-0cdb-48e7-a315-f2a3fdeee4df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.670034 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.670064 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.670077 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75wb8\" (UniqueName: \"kubernetes.io/projected/6da3d717-99fa-4202-909e-45db02ba3385-kube-api-access-75wb8\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.670092 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.670103 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdbfx\" (UniqueName: \"kubernetes.io/projected/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-kube-api-access-jdbfx\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.758477 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.768327 4715 generic.go:334] "Generic (PLEG): container finished" podID="6da3d717-99fa-4202-909e-45db02ba3385" containerID="fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56" exitCode=0 Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.768382 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bpdr" event={"ID":"6da3d717-99fa-4202-909e-45db02ba3385","Type":"ContainerDied","Data":"fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.768408 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bpdr" event={"ID":"6da3d717-99fa-4202-909e-45db02ba3385","Type":"ContainerDied","Data":"0ec3964531fe1b70f1974e66a2d6f6c654016b8ea416c059ff4baf4b17097c98"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.768425 4715 scope.go:117] "RemoveContainer" containerID="fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.768532 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2bpdr" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.771879 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m572w\" (UniqueName: \"kubernetes.io/projected/6467e871-dd98-4bcf-bf7c-53726589346b-kube-api-access-m572w\") pod \"6467e871-dd98-4bcf-bf7c-53726589346b\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.771927 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-trusted-ca\") pod \"6467e871-dd98-4bcf-bf7c-53726589346b\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.771971 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-operator-metrics\") pod \"6467e871-dd98-4bcf-bf7c-53726589346b\" (UID: \"6467e871-dd98-4bcf-bf7c-53726589346b\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.772632 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "6467e871-dd98-4bcf-bf7c-53726589346b" (UID: "6467e871-dd98-4bcf-bf7c-53726589346b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.782395 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6da3d717-99fa-4202-909e-45db02ba3385" (UID: "6da3d717-99fa-4202-909e-45db02ba3385"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.783919 4715 generic.go:334] "Generic (PLEG): container finished" podID="6af7e329-2107-4548-b087-77b8b60f6646" containerID="c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519" exitCode=0 Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.784670 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdrl9" event={"ID":"6af7e329-2107-4548-b087-77b8b60f6646","Type":"ContainerDied","Data":"c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.784743 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdrl9" event={"ID":"6af7e329-2107-4548-b087-77b8b60f6646","Type":"ContainerDied","Data":"6b3811d4d2322fb37501b2c0fbf2084e3b1fa3365e5eb3e0d32ec00fd8bbd140"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.787377 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdrl9" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.790327 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "6467e871-dd98-4bcf-bf7c-53726589346b" (UID: "6467e871-dd98-4bcf-bf7c-53726589346b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.790630 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6467e871-dd98-4bcf-bf7c-53726589346b-kube-api-access-m572w" (OuterVolumeSpecName: "kube-api-access-m572w") pod "6467e871-dd98-4bcf-bf7c-53726589346b" (UID: "6467e871-dd98-4bcf-bf7c-53726589346b"). InnerVolumeSpecName "kube-api-access-m572w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.794878 4715 generic.go:334] "Generic (PLEG): container finished" podID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerID="87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65" exitCode=0 Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.794976 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwc6v" event={"ID":"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df","Type":"ContainerDied","Data":"87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.795026 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwc6v" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.795039 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwc6v" event={"ID":"55c59d7a-0cdb-48e7-a315-f2a3fdeee4df","Type":"ContainerDied","Data":"1eba42764eb2367c0e2203b65b900874dc252827250bac4d71afb8b9eba1a9be"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.801662 4715 generic.go:334] "Generic (PLEG): container finished" podID="a3a35017-fd36-43de-83bd-8787fece173a" containerID="97d2f720f03fff20718215ecdc9f08bbac2ef1bf54d79efc33731ba620e1b706" exitCode=0 Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.801720 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxjdm" event={"ID":"a3a35017-fd36-43de-83bd-8787fece173a","Type":"ContainerDied","Data":"97d2f720f03fff20718215ecdc9f08bbac2ef1bf54d79efc33731ba620e1b706"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.824975 4715 generic.go:334] "Generic (PLEG): container finished" podID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerID="5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844" exitCode=0 Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.825124 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2nhl" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.825475 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2nhl" event={"ID":"cc4608d5-28b1-4b1a-94da-72f4f7bb7907","Type":"ContainerDied","Data":"5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.825502 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2nhl" event={"ID":"cc4608d5-28b1-4b1a-94da-72f4f7bb7907","Type":"ContainerDied","Data":"bf24d5bd240f6f5801e5f30a2d202e40e38ef3c82c7e390e975576afa34f11bf"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.834493 4715 generic.go:334] "Generic (PLEG): container finished" podID="6467e871-dd98-4bcf-bf7c-53726589346b" containerID="edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303" exitCode=0 Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.834606 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" event={"ID":"6467e871-dd98-4bcf-bf7c-53726589346b","Type":"ContainerDied","Data":"edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.834637 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" event={"ID":"6467e871-dd98-4bcf-bf7c-53726589346b","Type":"ContainerDied","Data":"c9a264aae6a04008ac0b82b97a804163fb63cd6646481d94e62aeaeb9d9a9ce9"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.834664 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2rstj" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.853388 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rgjtn" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerName="registry-server" containerID="cri-o://7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17" gracePeriod=30 Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.853613 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-25kht" event={"ID":"5c6da464-9fba-4591-a5b1-dfd37dde0770","Type":"ContainerStarted","Data":"d6644f9365cc8076671fa4fac1b0138c4e4c18b96c2962bfceb0353ed9629779"} Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.859114 4715 scope.go:117] "RemoveContainer" containerID="10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.870854 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwc6v"] Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.873453 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjv22\" (UniqueName: \"kubernetes.io/projected/6af7e329-2107-4548-b087-77b8b60f6646-kube-api-access-vjv22\") pod \"6af7e329-2107-4548-b087-77b8b60f6646\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.873529 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-utilities\") pod \"6af7e329-2107-4548-b087-77b8b60f6646\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.873572 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-catalog-content\") pod \"6af7e329-2107-4548-b087-77b8b60f6646\" (UID: \"6af7e329-2107-4548-b087-77b8b60f6646\") " Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.873806 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6da3d717-99fa-4202-909e-45db02ba3385-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.873816 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.873826 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m572w\" (UniqueName: \"kubernetes.io/projected/6467e871-dd98-4bcf-bf7c-53726589346b-kube-api-access-m572w\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.873835 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6467e871-dd98-4bcf-bf7c-53726589346b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.876140 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-utilities" (OuterVolumeSpecName: "utilities") pod "6af7e329-2107-4548-b087-77b8b60f6646" (UID: "6af7e329-2107-4548-b087-77b8b60f6646"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.877382 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6af7e329-2107-4548-b087-77b8b60f6646-kube-api-access-vjv22" (OuterVolumeSpecName: "kube-api-access-vjv22") pod "6af7e329-2107-4548-b087-77b8b60f6646" (UID: "6af7e329-2107-4548-b087-77b8b60f6646"). InnerVolumeSpecName "kube-api-access-vjv22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.878286 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwc6v"] Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.896466 4715 scope.go:117] "RemoveContainer" containerID="033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.905330 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rstj"] Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.910817 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc4608d5-28b1-4b1a-94da-72f4f7bb7907" (UID: "cc4608d5-28b1-4b1a-94da-72f4f7bb7907"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.914749 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rstj"] Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.924200 4715 scope.go:117] "RemoveContainer" containerID="fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56" Nov 25 12:13:14 crc kubenswrapper[4715]: E1125 12:13:14.924656 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56\": container with ID starting with fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56 not found: ID does not exist" containerID="fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.924690 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56"} err="failed to get container status \"fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56\": rpc error: code = NotFound desc = could not find container \"fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56\": container with ID starting with fe9a0bfecb3890567c4ad20c8031d13978e4a26e043d94fa5792b71430f77c56 not found: ID does not exist" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.924712 4715 scope.go:117] "RemoveContainer" containerID="10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4" Nov 25 12:13:14 crc kubenswrapper[4715]: E1125 12:13:14.926071 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4\": container with ID starting with 10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4 not found: ID does not exist" containerID="10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.926519 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4"} err="failed to get container status \"10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4\": rpc error: code = NotFound desc = could not find container \"10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4\": container with ID starting with 10b198ea894a7494795d348d4cb1552ce8ec7c70a1bf19be7b81fb8bae992fa4 not found: ID does not exist" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.926544 4715 scope.go:117] "RemoveContainer" containerID="033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8" Nov 25 12:13:14 crc kubenswrapper[4715]: E1125 12:13:14.926922 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8\": container with ID starting with 033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8 not found: ID does not exist" containerID="033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.926977 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8"} err="failed to get container status \"033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8\": rpc error: code = NotFound desc = could not find container \"033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8\": container with ID starting with 033a4df411fb1fee1dff7d81f6e60f62306f56e45e33f904c8250761b435feb8 not found: ID does not exist" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.927013 4715 scope.go:117] "RemoveContainer" containerID="c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.946032 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6af7e329-2107-4548-b087-77b8b60f6646" (UID: "6af7e329-2107-4548-b087-77b8b60f6646"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.948937 4715 scope.go:117] "RemoveContainer" containerID="1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.975306 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjv22\" (UniqueName: \"kubernetes.io/projected/6af7e329-2107-4548-b087-77b8b60f6646-kube-api-access-vjv22\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.975345 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.975361 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4608d5-28b1-4b1a-94da-72f4f7bb7907-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.975374 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6af7e329-2107-4548-b087-77b8b60f6646-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.978362 4715 scope.go:117] "RemoveContainer" containerID="4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.992710 4715 scope.go:117] "RemoveContainer" containerID="c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519" Nov 25 12:13:14 crc kubenswrapper[4715]: E1125 12:13:14.993503 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519\": container with ID starting with c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519 not found: ID does not exist" containerID="c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.993551 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519"} err="failed to get container status \"c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519\": rpc error: code = NotFound desc = could not find container \"c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519\": container with ID starting with c0dd67b09efd747a96e76602b612e99fcce54a2db26c78d951f630e80cb6f519 not found: ID does not exist" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.993589 4715 scope.go:117] "RemoveContainer" containerID="1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43" Nov 25 12:13:14 crc kubenswrapper[4715]: E1125 12:13:14.993988 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43\": container with ID starting with 1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43 not found: ID does not exist" containerID="1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.994018 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43"} err="failed to get container status \"1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43\": rpc error: code = NotFound desc = could not find container \"1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43\": container with ID starting with 1e4fd90f9c0cafaef9942df59f958a3da61ac413337fae55a8dcff85225dcf43 not found: ID does not exist" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.994046 4715 scope.go:117] "RemoveContainer" containerID="4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b" Nov 25 12:13:14 crc kubenswrapper[4715]: E1125 12:13:14.994563 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b\": container with ID starting with 4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b not found: ID does not exist" containerID="4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.994589 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b"} err="failed to get container status \"4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b\": rpc error: code = NotFound desc = could not find container \"4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b\": container with ID starting with 4662334a9dd6507ed2d88f8d3cb193b94636b09f389de706c5df1c9ff279836b not found: ID does not exist" Nov 25 12:13:14 crc kubenswrapper[4715]: I1125 12:13:14.994606 4715 scope.go:117] "RemoveContainer" containerID="87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.006624 4715 scope.go:117] "RemoveContainer" containerID="cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.022280 4715 scope.go:117] "RemoveContainer" containerID="2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.035423 4715 scope.go:117] "RemoveContainer" containerID="87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.035976 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65\": container with ID starting with 87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65 not found: ID does not exist" containerID="87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.036021 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65"} err="failed to get container status \"87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65\": rpc error: code = NotFound desc = could not find container \"87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65\": container with ID starting with 87035ad05185594ebc8285dca4fe0a6244f393e1035be265f002fd3d06a95a65 not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.036044 4715 scope.go:117] "RemoveContainer" containerID="cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.036380 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31\": container with ID starting with cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31 not found: ID does not exist" containerID="cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.036405 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31"} err="failed to get container status \"cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31\": rpc error: code = NotFound desc = could not find container \"cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31\": container with ID starting with cf04096155fa4e10007b77f1cb6fe57de11f2b48701099fa7d6daec6fda84b31 not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.036424 4715 scope.go:117] "RemoveContainer" containerID="2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.036665 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf\": container with ID starting with 2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf not found: ID does not exist" containerID="2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.036687 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf"} err="failed to get container status \"2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf\": rpc error: code = NotFound desc = could not find container \"2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf\": container with ID starting with 2466640a4df1a96c34bab8e6e7f67ecf6948c6ca2741bc196c873ffb334953cf not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.036701 4715 scope.go:117] "RemoveContainer" containerID="5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.051055 4715 scope.go:117] "RemoveContainer" containerID="fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.066777 4715 scope.go:117] "RemoveContainer" containerID="f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.088465 4715 scope.go:117] "RemoveContainer" containerID="5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.088868 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844\": container with ID starting with 5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844 not found: ID does not exist" containerID="5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.088910 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844"} err="failed to get container status \"5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844\": rpc error: code = NotFound desc = could not find container \"5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844\": container with ID starting with 5be185a16d69fa53d1fa20883e876f8dd11102b9ee2b7260af05d6191300a844 not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.088948 4715 scope.go:117] "RemoveContainer" containerID="fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.089647 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d\": container with ID starting with fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d not found: ID does not exist" containerID="fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.089690 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d"} err="failed to get container status \"fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d\": rpc error: code = NotFound desc = could not find container \"fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d\": container with ID starting with fbcf0ba456d870fed6f9bb2bb7302b2ce4ae040f9425676b35bea214da0b7f3d not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.089718 4715 scope.go:117] "RemoveContainer" containerID="f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.090073 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1\": container with ID starting with f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1 not found: ID does not exist" containerID="f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.090108 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1"} err="failed to get container status \"f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1\": rpc error: code = NotFound desc = could not find container \"f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1\": container with ID starting with f8a30d5617fac0c528622a443511f0cae9f2f45a8b3e28d66be42e2839d3fea1 not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.090128 4715 scope.go:117] "RemoveContainer" containerID="edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.098711 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2bpdr"] Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.106018 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2bpdr"] Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.109924 4715 scope.go:117] "RemoveContainer" containerID="edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.110565 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303\": container with ID starting with edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303 not found: ID does not exist" containerID="edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.110597 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303"} err="failed to get container status \"edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303\": rpc error: code = NotFound desc = could not find container \"edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303\": container with ID starting with edba134bde5511b02debad2c820532d1f33ec5a7a35db5c66059dbdf9dfd6303 not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.116040 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gdrl9"] Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.123823 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gdrl9"] Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.160263 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2nhl"] Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.165422 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z2nhl"] Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.267612 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.379624 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-catalog-content\") pod \"a3a35017-fd36-43de-83bd-8787fece173a\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.379769 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r55px\" (UniqueName: \"kubernetes.io/projected/a3a35017-fd36-43de-83bd-8787fece173a-kube-api-access-r55px\") pod \"a3a35017-fd36-43de-83bd-8787fece173a\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.379811 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-utilities\") pod \"a3a35017-fd36-43de-83bd-8787fece173a\" (UID: \"a3a35017-fd36-43de-83bd-8787fece173a\") " Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.381129 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-utilities" (OuterVolumeSpecName: "utilities") pod "a3a35017-fd36-43de-83bd-8787fece173a" (UID: "a3a35017-fd36-43de-83bd-8787fece173a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.385261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a35017-fd36-43de-83bd-8787fece173a-kube-api-access-r55px" (OuterVolumeSpecName: "kube-api-access-r55px") pod "a3a35017-fd36-43de-83bd-8787fece173a" (UID: "a3a35017-fd36-43de-83bd-8787fece173a"). InnerVolumeSpecName "kube-api-access-r55px". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.481470 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r55px\" (UniqueName: \"kubernetes.io/projected/a3a35017-fd36-43de-83bd-8787fece173a-kube-api-access-r55px\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.481503 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.599770 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3a35017-fd36-43de-83bd-8787fece173a" (UID: "a3a35017-fd36-43de-83bd-8787fece173a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.684023 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3a35017-fd36-43de-83bd-8787fece173a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.837332 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.862099 4715 generic.go:334] "Generic (PLEG): container finished" podID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerID="7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17" exitCode=0 Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.862155 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rgjtn" event={"ID":"b3d2d042-e214-44ff-9c25-9d894bab99f3","Type":"ContainerDied","Data":"7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17"} Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.862179 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rgjtn" event={"ID":"b3d2d042-e214-44ff-9c25-9d894bab99f3","Type":"ContainerDied","Data":"a6c5a6bbff0b28e073b9b5f88d096caae3451daf427ec89c362601763dcb243b"} Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.862213 4715 scope.go:117] "RemoveContainer" containerID="7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.862316 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rgjtn" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.863989 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-25kht" event={"ID":"5c6da464-9fba-4591-a5b1-dfd37dde0770","Type":"ContainerStarted","Data":"7d04cef5b5fc5d1cc9df712be3096f9c32f5043db99a748e0d5d79d8376c67ae"} Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.865611 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.870098 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-25kht" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.874253 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxjdm" event={"ID":"a3a35017-fd36-43de-83bd-8787fece173a","Type":"ContainerDied","Data":"457d5b985d3acc4616241e0b8ff98a65e1f200fe38983a1a15f401e97aa34a97"} Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.874340 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxjdm" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.887064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4k5w\" (UniqueName: \"kubernetes.io/projected/b3d2d042-e214-44ff-9c25-9d894bab99f3-kube-api-access-b4k5w\") pod \"b3d2d042-e214-44ff-9c25-9d894bab99f3\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.887123 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-utilities\") pod \"b3d2d042-e214-44ff-9c25-9d894bab99f3\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.887176 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-catalog-content\") pod \"b3d2d042-e214-44ff-9c25-9d894bab99f3\" (UID: \"b3d2d042-e214-44ff-9c25-9d894bab99f3\") " Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.890374 4715 scope.go:117] "RemoveContainer" containerID="1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.895226 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-utilities" (OuterVolumeSpecName: "utilities") pod "b3d2d042-e214-44ff-9c25-9d894bab99f3" (UID: "b3d2d042-e214-44ff-9c25-9d894bab99f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.898962 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d2d042-e214-44ff-9c25-9d894bab99f3-kube-api-access-b4k5w" (OuterVolumeSpecName: "kube-api-access-b4k5w") pod "b3d2d042-e214-44ff-9c25-9d894bab99f3" (UID: "b3d2d042-e214-44ff-9c25-9d894bab99f3"). InnerVolumeSpecName "kube-api-access-b4k5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.921174 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-25kht" podStartSLOduration=2.9211544739999997 podStartE2EDuration="2.921154474s" podCreationTimestamp="2025-11-25 12:13:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:13:15.890536993 +0000 UTC m=+246.398040004" watchObservedRunningTime="2025-11-25 12:13:15.921154474 +0000 UTC m=+246.428657495" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.925305 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3d2d042-e214-44ff-9c25-9d894bab99f3" (UID: "b3d2d042-e214-44ff-9c25-9d894bab99f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.926036 4715 scope.go:117] "RemoveContainer" containerID="634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.941648 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zxjdm"] Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.944879 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zxjdm"] Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.948790 4715 scope.go:117] "RemoveContainer" containerID="7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.951012 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17\": container with ID starting with 7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17 not found: ID does not exist" containerID="7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.951062 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17"} err="failed to get container status \"7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17\": rpc error: code = NotFound desc = could not find container \"7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17\": container with ID starting with 7676ae97d8f459ce625ad5b6d2b1b65ad482339d6f06508ab9aa55010262bc17 not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.951092 4715 scope.go:117] "RemoveContainer" containerID="1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.951758 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9\": container with ID starting with 1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9 not found: ID does not exist" containerID="1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.951782 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9"} err="failed to get container status \"1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9\": rpc error: code = NotFound desc = could not find container \"1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9\": container with ID starting with 1faa6a51ed5c6c4027e4a515ac2ed213420fc9145e684b1ac4e80a45e4cb77e9 not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.951795 4715 scope.go:117] "RemoveContainer" containerID="634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d" Nov 25 12:13:15 crc kubenswrapper[4715]: E1125 12:13:15.952090 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d\": container with ID starting with 634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d not found: ID does not exist" containerID="634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.952111 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d"} err="failed to get container status \"634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d\": rpc error: code = NotFound desc = could not find container \"634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d\": container with ID starting with 634ed916e5956dab5420ab71d0fd2a0881b8697ce2d68ca68f63b0d112c7ff2d not found: ID does not exist" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.952124 4715 scope.go:117] "RemoveContainer" containerID="97d2f720f03fff20718215ecdc9f08bbac2ef1bf54d79efc33731ba620e1b706" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.972162 4715 scope.go:117] "RemoveContainer" containerID="32759f7febe258537e90b6193ea71ca32d0566c5dd5dd7428d6cd87ea73552ef" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.988384 4715 scope.go:117] "RemoveContainer" containerID="89e1f5578d926e6633f83eb4ff0a5f6e8481a67d17a19f11565ec7fa60fe867b" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.988827 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4k5w\" (UniqueName: \"kubernetes.io/projected/b3d2d042-e214-44ff-9c25-9d894bab99f3-kube-api-access-b4k5w\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.988847 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:15 crc kubenswrapper[4715]: I1125 12:13:15.988860 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d042-e214-44ff-9c25-9d894bab99f3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:16 crc kubenswrapper[4715]: I1125 12:13:16.198664 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rgjtn"] Nov 25 12:13:16 crc kubenswrapper[4715]: I1125 12:13:16.201939 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rgjtn"] Nov 25 12:13:16 crc kubenswrapper[4715]: I1125 12:13:16.703654 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" path="/var/lib/kubelet/pods/55c59d7a-0cdb-48e7-a315-f2a3fdeee4df/volumes" Nov 25 12:13:16 crc kubenswrapper[4715]: I1125 12:13:16.704605 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6467e871-dd98-4bcf-bf7c-53726589346b" path="/var/lib/kubelet/pods/6467e871-dd98-4bcf-bf7c-53726589346b/volumes" Nov 25 12:13:16 crc kubenswrapper[4715]: I1125 12:13:16.705291 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6af7e329-2107-4548-b087-77b8b60f6646" path="/var/lib/kubelet/pods/6af7e329-2107-4548-b087-77b8b60f6646/volumes" Nov 25 12:13:16 crc kubenswrapper[4715]: I1125 12:13:16.706659 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da3d717-99fa-4202-909e-45db02ba3385" path="/var/lib/kubelet/pods/6da3d717-99fa-4202-909e-45db02ba3385/volumes" Nov 25 12:13:16 crc kubenswrapper[4715]: I1125 12:13:16.707589 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3a35017-fd36-43de-83bd-8787fece173a" path="/var/lib/kubelet/pods/a3a35017-fd36-43de-83bd-8787fece173a/volumes" Nov 25 12:13:16 crc kubenswrapper[4715]: I1125 12:13:16.709030 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" path="/var/lib/kubelet/pods/b3d2d042-e214-44ff-9c25-9d894bab99f3/volumes" Nov 25 12:13:16 crc kubenswrapper[4715]: I1125 12:13:16.709981 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" path="/var/lib/kubelet/pods/cc4608d5-28b1-4b1a-94da-72f4f7bb7907/volumes" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.170214 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zp575"] Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.171933 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6af7e329-2107-4548-b087-77b8b60f6646" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.172039 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6af7e329-2107-4548-b087-77b8b60f6646" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.172327 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.172443 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.172640 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da3d717-99fa-4202-909e-45db02ba3385" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.172727 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da3d717-99fa-4202-909e-45db02ba3385" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.172814 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.172942 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.173029 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da3d717-99fa-4202-909e-45db02ba3385" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.173113 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da3d717-99fa-4202-909e-45db02ba3385" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.173222 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.173311 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.175153 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6af7e329-2107-4548-b087-77b8b60f6646" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.175402 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6af7e329-2107-4548-b087-77b8b60f6646" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.175510 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6467e871-dd98-4bcf-bf7c-53726589346b" containerName="marketplace-operator" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.175593 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6467e871-dd98-4bcf-bf7c-53726589346b" containerName="marketplace-operator" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.175670 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da3d717-99fa-4202-909e-45db02ba3385" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.175757 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da3d717-99fa-4202-909e-45db02ba3385" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.175834 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a35017-fd36-43de-83bd-8787fece173a" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.175912 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a35017-fd36-43de-83bd-8787fece173a" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.176053 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.178511 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.178637 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.178730 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.178815 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a35017-fd36-43de-83bd-8787fece173a" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.178893 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a35017-fd36-43de-83bd-8787fece173a" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.178984 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.179062 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.179140 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.179245 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.179339 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6af7e329-2107-4548-b087-77b8b60f6646" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.179415 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6af7e329-2107-4548-b087-77b8b60f6646" containerName="extract-utilities" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.179493 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.179567 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.179647 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a35017-fd36-43de-83bd-8787fece173a" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.179732 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a35017-fd36-43de-83bd-8787fece173a" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: E1125 12:13:17.179815 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.179892 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerName="extract-content" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.180229 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a35017-fd36-43de-83bd-8787fece173a" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.180337 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6af7e329-2107-4548-b087-77b8b60f6646" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.180420 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6467e871-dd98-4bcf-bf7c-53726589346b" containerName="marketplace-operator" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.180530 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da3d717-99fa-4202-909e-45db02ba3385" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.180614 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d2d042-e214-44ff-9c25-9d894bab99f3" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.180700 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc4608d5-28b1-4b1a-94da-72f4f7bb7907" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.180777 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="55c59d7a-0cdb-48e7-a315-f2a3fdeee4df" containerName="registry-server" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.181760 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.182341 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zp575"] Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.183986 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.308142 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4dbg\" (UniqueName: \"kubernetes.io/projected/c4034c63-b17b-4799-a34e-4432e780c343-kube-api-access-v4dbg\") pod \"community-operators-zp575\" (UID: \"c4034c63-b17b-4799-a34e-4432e780c343\") " pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.308243 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4034c63-b17b-4799-a34e-4432e780c343-catalog-content\") pod \"community-operators-zp575\" (UID: \"c4034c63-b17b-4799-a34e-4432e780c343\") " pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.308342 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4034c63-b17b-4799-a34e-4432e780c343-utilities\") pod \"community-operators-zp575\" (UID: \"c4034c63-b17b-4799-a34e-4432e780c343\") " pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.409514 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4034c63-b17b-4799-a34e-4432e780c343-catalog-content\") pod \"community-operators-zp575\" (UID: \"c4034c63-b17b-4799-a34e-4432e780c343\") " pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.409627 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4034c63-b17b-4799-a34e-4432e780c343-utilities\") pod \"community-operators-zp575\" (UID: \"c4034c63-b17b-4799-a34e-4432e780c343\") " pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.409691 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4dbg\" (UniqueName: \"kubernetes.io/projected/c4034c63-b17b-4799-a34e-4432e780c343-kube-api-access-v4dbg\") pod \"community-operators-zp575\" (UID: \"c4034c63-b17b-4799-a34e-4432e780c343\") " pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.410105 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4034c63-b17b-4799-a34e-4432e780c343-catalog-content\") pod \"community-operators-zp575\" (UID: \"c4034c63-b17b-4799-a34e-4432e780c343\") " pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.410118 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4034c63-b17b-4799-a34e-4432e780c343-utilities\") pod \"community-operators-zp575\" (UID: \"c4034c63-b17b-4799-a34e-4432e780c343\") " pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.427836 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4dbg\" (UniqueName: \"kubernetes.io/projected/c4034c63-b17b-4799-a34e-4432e780c343-kube-api-access-v4dbg\") pod \"community-operators-zp575\" (UID: \"c4034c63-b17b-4799-a34e-4432e780c343\") " pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.510761 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.771636 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w95rs"] Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.773071 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.777208 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.780118 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w95rs"] Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.817151 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf866d85-2d8e-4ea2-b1b8-4cd1001beb74-utilities\") pod \"redhat-operators-w95rs\" (UID: \"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74\") " pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.817267 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j8pz\" (UniqueName: \"kubernetes.io/projected/bf866d85-2d8e-4ea2-b1b8-4cd1001beb74-kube-api-access-9j8pz\") pod \"redhat-operators-w95rs\" (UID: \"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74\") " pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.817381 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf866d85-2d8e-4ea2-b1b8-4cd1001beb74-catalog-content\") pod \"redhat-operators-w95rs\" (UID: \"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74\") " pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.914051 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zp575"] Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.918049 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf866d85-2d8e-4ea2-b1b8-4cd1001beb74-utilities\") pod \"redhat-operators-w95rs\" (UID: \"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74\") " pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.918103 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j8pz\" (UniqueName: \"kubernetes.io/projected/bf866d85-2d8e-4ea2-b1b8-4cd1001beb74-kube-api-access-9j8pz\") pod \"redhat-operators-w95rs\" (UID: \"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74\") " pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.918155 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf866d85-2d8e-4ea2-b1b8-4cd1001beb74-catalog-content\") pod \"redhat-operators-w95rs\" (UID: \"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74\") " pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.918588 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf866d85-2d8e-4ea2-b1b8-4cd1001beb74-catalog-content\") pod \"redhat-operators-w95rs\" (UID: \"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74\") " pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.918699 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf866d85-2d8e-4ea2-b1b8-4cd1001beb74-utilities\") pod \"redhat-operators-w95rs\" (UID: \"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74\") " pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:17 crc kubenswrapper[4715]: W1125 12:13:17.926005 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4034c63_b17b_4799_a34e_4432e780c343.slice/crio-1384bf5dd17841ee34d92800ec84b4002638c094a9ae2299a31e4fc862d61a17 WatchSource:0}: Error finding container 1384bf5dd17841ee34d92800ec84b4002638c094a9ae2299a31e4fc862d61a17: Status 404 returned error can't find the container with id 1384bf5dd17841ee34d92800ec84b4002638c094a9ae2299a31e4fc862d61a17 Nov 25 12:13:17 crc kubenswrapper[4715]: I1125 12:13:17.936025 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j8pz\" (UniqueName: \"kubernetes.io/projected/bf866d85-2d8e-4ea2-b1b8-4cd1001beb74-kube-api-access-9j8pz\") pod \"redhat-operators-w95rs\" (UID: \"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74\") " pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:18 crc kubenswrapper[4715]: I1125 12:13:18.093436 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:18 crc kubenswrapper[4715]: I1125 12:13:18.288645 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w95rs"] Nov 25 12:13:18 crc kubenswrapper[4715]: I1125 12:13:18.917719 4715 generic.go:334] "Generic (PLEG): container finished" podID="c4034c63-b17b-4799-a34e-4432e780c343" containerID="8d51c29ef499a2b47552620e917a35d73d76471cf95e30e76b4eb252a9062e23" exitCode=0 Nov 25 12:13:18 crc kubenswrapper[4715]: I1125 12:13:18.917825 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp575" event={"ID":"c4034c63-b17b-4799-a34e-4432e780c343","Type":"ContainerDied","Data":"8d51c29ef499a2b47552620e917a35d73d76471cf95e30e76b4eb252a9062e23"} Nov 25 12:13:18 crc kubenswrapper[4715]: I1125 12:13:18.917871 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp575" event={"ID":"c4034c63-b17b-4799-a34e-4432e780c343","Type":"ContainerStarted","Data":"1384bf5dd17841ee34d92800ec84b4002638c094a9ae2299a31e4fc862d61a17"} Nov 25 12:13:18 crc kubenswrapper[4715]: I1125 12:13:18.919844 4715 generic.go:334] "Generic (PLEG): container finished" podID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" containerID="94924341cb027faf16a0776eb1a7a4efa1bd0d20b4a01f51755c778e77ce7984" exitCode=0 Nov 25 12:13:18 crc kubenswrapper[4715]: I1125 12:13:18.919877 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w95rs" event={"ID":"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74","Type":"ContainerDied","Data":"94924341cb027faf16a0776eb1a7a4efa1bd0d20b4a01f51755c778e77ce7984"} Nov 25 12:13:18 crc kubenswrapper[4715]: I1125 12:13:18.919899 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w95rs" event={"ID":"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74","Type":"ContainerStarted","Data":"8ca804fe9d1aedda5f183e9ed28d40e504bc7c85f62532b8efab2f4efef36f20"} Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.573532 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xjzs4"] Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.574770 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.580622 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.583525 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xjzs4"] Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.637912 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6590884-9e6f-4390-b4c1-59902c2684ac-utilities\") pod \"certified-operators-xjzs4\" (UID: \"b6590884-9e6f-4390-b4c1-59902c2684ac\") " pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.637983 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd95w\" (UniqueName: \"kubernetes.io/projected/b6590884-9e6f-4390-b4c1-59902c2684ac-kube-api-access-pd95w\") pod \"certified-operators-xjzs4\" (UID: \"b6590884-9e6f-4390-b4c1-59902c2684ac\") " pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.638013 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6590884-9e6f-4390-b4c1-59902c2684ac-catalog-content\") pod \"certified-operators-xjzs4\" (UID: \"b6590884-9e6f-4390-b4c1-59902c2684ac\") " pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.739571 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd95w\" (UniqueName: \"kubernetes.io/projected/b6590884-9e6f-4390-b4c1-59902c2684ac-kube-api-access-pd95w\") pod \"certified-operators-xjzs4\" (UID: \"b6590884-9e6f-4390-b4c1-59902c2684ac\") " pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.739648 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6590884-9e6f-4390-b4c1-59902c2684ac-catalog-content\") pod \"certified-operators-xjzs4\" (UID: \"b6590884-9e6f-4390-b4c1-59902c2684ac\") " pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.739724 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6590884-9e6f-4390-b4c1-59902c2684ac-utilities\") pod \"certified-operators-xjzs4\" (UID: \"b6590884-9e6f-4390-b4c1-59902c2684ac\") " pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.740211 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6590884-9e6f-4390-b4c1-59902c2684ac-utilities\") pod \"certified-operators-xjzs4\" (UID: \"b6590884-9e6f-4390-b4c1-59902c2684ac\") " pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.740286 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6590884-9e6f-4390-b4c1-59902c2684ac-catalog-content\") pod \"certified-operators-xjzs4\" (UID: \"b6590884-9e6f-4390-b4c1-59902c2684ac\") " pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.760492 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd95w\" (UniqueName: \"kubernetes.io/projected/b6590884-9e6f-4390-b4c1-59902c2684ac-kube-api-access-pd95w\") pod \"certified-operators-xjzs4\" (UID: \"b6590884-9e6f-4390-b4c1-59902c2684ac\") " pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.903970 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:19 crc kubenswrapper[4715]: I1125 12:13:19.929643 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp575" event={"ID":"c4034c63-b17b-4799-a34e-4432e780c343","Type":"ContainerStarted","Data":"a11cb80be06ed8de35f6b9eb1cdd5dc26662d4c980ab5edaa8a9b083f14ec951"} Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.175118 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-thpk4"] Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.176841 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.179599 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.191792 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-thpk4"] Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.245571 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44cf8567-083d-44d6-bab3-a90ee6497454-utilities\") pod \"redhat-marketplace-thpk4\" (UID: \"44cf8567-083d-44d6-bab3-a90ee6497454\") " pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.245647 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g8v5\" (UniqueName: \"kubernetes.io/projected/44cf8567-083d-44d6-bab3-a90ee6497454-kube-api-access-9g8v5\") pod \"redhat-marketplace-thpk4\" (UID: \"44cf8567-083d-44d6-bab3-a90ee6497454\") " pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.245729 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44cf8567-083d-44d6-bab3-a90ee6497454-catalog-content\") pod \"redhat-marketplace-thpk4\" (UID: \"44cf8567-083d-44d6-bab3-a90ee6497454\") " pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.288892 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xjzs4"] Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.347348 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44cf8567-083d-44d6-bab3-a90ee6497454-utilities\") pod \"redhat-marketplace-thpk4\" (UID: \"44cf8567-083d-44d6-bab3-a90ee6497454\") " pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.347440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g8v5\" (UniqueName: \"kubernetes.io/projected/44cf8567-083d-44d6-bab3-a90ee6497454-kube-api-access-9g8v5\") pod \"redhat-marketplace-thpk4\" (UID: \"44cf8567-083d-44d6-bab3-a90ee6497454\") " pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.347498 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44cf8567-083d-44d6-bab3-a90ee6497454-catalog-content\") pod \"redhat-marketplace-thpk4\" (UID: \"44cf8567-083d-44d6-bab3-a90ee6497454\") " pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.347823 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44cf8567-083d-44d6-bab3-a90ee6497454-utilities\") pod \"redhat-marketplace-thpk4\" (UID: \"44cf8567-083d-44d6-bab3-a90ee6497454\") " pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.347909 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44cf8567-083d-44d6-bab3-a90ee6497454-catalog-content\") pod \"redhat-marketplace-thpk4\" (UID: \"44cf8567-083d-44d6-bab3-a90ee6497454\") " pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.371494 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g8v5\" (UniqueName: \"kubernetes.io/projected/44cf8567-083d-44d6-bab3-a90ee6497454-kube-api-access-9g8v5\") pod \"redhat-marketplace-thpk4\" (UID: \"44cf8567-083d-44d6-bab3-a90ee6497454\") " pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.522111 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.906950 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-thpk4"] Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.950109 4715 generic.go:334] "Generic (PLEG): container finished" podID="c4034c63-b17b-4799-a34e-4432e780c343" containerID="a11cb80be06ed8de35f6b9eb1cdd5dc26662d4c980ab5edaa8a9b083f14ec951" exitCode=0 Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.950230 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp575" event={"ID":"c4034c63-b17b-4799-a34e-4432e780c343","Type":"ContainerDied","Data":"a11cb80be06ed8de35f6b9eb1cdd5dc26662d4c980ab5edaa8a9b083f14ec951"} Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.955895 4715 generic.go:334] "Generic (PLEG): container finished" podID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" containerID="e85ef7a657bab144d1a8e1620a1452a617aa7576e224f790404953b1e65023d3" exitCode=0 Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.956089 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w95rs" event={"ID":"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74","Type":"ContainerDied","Data":"e85ef7a657bab144d1a8e1620a1452a617aa7576e224f790404953b1e65023d3"} Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.959605 4715 generic.go:334] "Generic (PLEG): container finished" podID="b6590884-9e6f-4390-b4c1-59902c2684ac" containerID="7868f062704316e9957e683c09fc995213f926479b826432e0d48796532c2635" exitCode=0 Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.959706 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjzs4" event={"ID":"b6590884-9e6f-4390-b4c1-59902c2684ac","Type":"ContainerDied","Data":"7868f062704316e9957e683c09fc995213f926479b826432e0d48796532c2635"} Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.959732 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjzs4" event={"ID":"b6590884-9e6f-4390-b4c1-59902c2684ac","Type":"ContainerStarted","Data":"480cd6d5702f0929bf9b9e75aab849606c9c5e3dae63ecb06410a180613b5ebc"} Nov 25 12:13:20 crc kubenswrapper[4715]: I1125 12:13:20.972505 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thpk4" event={"ID":"44cf8567-083d-44d6-bab3-a90ee6497454","Type":"ContainerStarted","Data":"2c6e7924c0c39275670260ae7e8ce6c31cf5c0832447d5af0b28c98033a4c668"} Nov 25 12:13:21 crc kubenswrapper[4715]: I1125 12:13:21.978839 4715 generic.go:334] "Generic (PLEG): container finished" podID="44cf8567-083d-44d6-bab3-a90ee6497454" containerID="eade445606caf548b9bc887f7feb07e9b3156e9f706e4027d8349a1cfa6c1360" exitCode=0 Nov 25 12:13:21 crc kubenswrapper[4715]: I1125 12:13:21.978951 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thpk4" event={"ID":"44cf8567-083d-44d6-bab3-a90ee6497454","Type":"ContainerDied","Data":"eade445606caf548b9bc887f7feb07e9b3156e9f706e4027d8349a1cfa6c1360"} Nov 25 12:13:22 crc kubenswrapper[4715]: I1125 12:13:22.985991 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zp575" event={"ID":"c4034c63-b17b-4799-a34e-4432e780c343","Type":"ContainerStarted","Data":"ebd28f8b4cb70c5739ce4586eaef789bd375774bc344c8bdf2199a64235d6495"} Nov 25 12:13:22 crc kubenswrapper[4715]: I1125 12:13:22.988754 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w95rs" event={"ID":"bf866d85-2d8e-4ea2-b1b8-4cd1001beb74","Type":"ContainerStarted","Data":"518757fa7f13c23f44f6814a499888bfe1be01e47c7b439b5ec51a70816c6964"} Nov 25 12:13:22 crc kubenswrapper[4715]: I1125 12:13:22.990422 4715 generic.go:334] "Generic (PLEG): container finished" podID="b6590884-9e6f-4390-b4c1-59902c2684ac" containerID="3d6c924ad82777227e775e2e7ed460956b6c80e012c77cdfa1d80a0de9181c83" exitCode=0 Nov 25 12:13:22 crc kubenswrapper[4715]: I1125 12:13:22.990459 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjzs4" event={"ID":"b6590884-9e6f-4390-b4c1-59902c2684ac","Type":"ContainerDied","Data":"3d6c924ad82777227e775e2e7ed460956b6c80e012c77cdfa1d80a0de9181c83"} Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.003686 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zp575" podStartSLOduration=3.261181672 podStartE2EDuration="6.003670205s" podCreationTimestamp="2025-11-25 12:13:17 +0000 UTC" firstStartedPulling="2025-11-25 12:13:18.921044184 +0000 UTC m=+249.428547205" lastFinishedPulling="2025-11-25 12:13:21.663532717 +0000 UTC m=+252.171035738" observedRunningTime="2025-11-25 12:13:23.000379531 +0000 UTC m=+253.507882562" watchObservedRunningTime="2025-11-25 12:13:23.003670205 +0000 UTC m=+253.511173216" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.021843 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w95rs" podStartSLOduration=2.653969671 podStartE2EDuration="6.021821165s" podCreationTimestamp="2025-11-25 12:13:17 +0000 UTC" firstStartedPulling="2025-11-25 12:13:18.92125681 +0000 UTC m=+249.428759831" lastFinishedPulling="2025-11-25 12:13:22.289108304 +0000 UTC m=+252.796611325" observedRunningTime="2025-11-25 12:13:23.019569801 +0000 UTC m=+253.527072832" watchObservedRunningTime="2025-11-25 12:13:23.021821165 +0000 UTC m=+253.529324186" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.347103 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.347774 4715 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.347791 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.348150 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728" gracePeriod=15 Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.348210 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae" gracePeriod=15 Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.348213 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd" gracePeriod=15 Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.348245 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4" gracePeriod=15 Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.348128 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1" gracePeriod=15 Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.350036 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 12:13:23 crc kubenswrapper[4715]: E1125 12:13:23.351289 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.351418 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 12:13:23 crc kubenswrapper[4715]: E1125 12:13:23.351508 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.351642 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 12:13:23 crc kubenswrapper[4715]: E1125 12:13:23.351728 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.351827 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 12:13:23 crc kubenswrapper[4715]: E1125 12:13:23.351914 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.351994 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 12:13:23 crc kubenswrapper[4715]: E1125 12:13:23.352079 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.352243 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 12:13:23 crc kubenswrapper[4715]: E1125 12:13:23.352390 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.352486 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 12:13:23 crc kubenswrapper[4715]: E1125 12:13:23.352598 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.352804 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.353229 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.353316 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.353400 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.353483 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.353563 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.353646 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 12:13:23 crc kubenswrapper[4715]: E1125 12:13:23.353843 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.354083 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.354253 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.393087 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.393394 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.393502 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.393583 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.393661 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.393748 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.393880 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.393976 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.495835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496282 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496317 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496339 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496362 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496384 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496399 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496431 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496511 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496016 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496557 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496578 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496597 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496617 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496640 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.496663 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.998629 4715 generic.go:334] "Generic (PLEG): container finished" podID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" containerID="52fd2fe0a4942b9ac1d15de7547dc8b0d30b1c8f6fb2a82f8c0659d539f10576" exitCode=0 Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.998698 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2","Type":"ContainerDied","Data":"52fd2fe0a4942b9ac1d15de7547dc8b0d30b1c8f6fb2a82f8c0659d539f10576"} Nov 25 12:13:23 crc kubenswrapper[4715]: I1125 12:13:23.999772 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:24 crc kubenswrapper[4715]: I1125 12:13:24.000296 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:24 crc kubenswrapper[4715]: I1125 12:13:24.001415 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 12:13:24 crc kubenswrapper[4715]: I1125 12:13:24.002503 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 12:13:24 crc kubenswrapper[4715]: I1125 12:13:24.003208 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728" exitCode=0 Nov 25 12:13:24 crc kubenswrapper[4715]: I1125 12:13:24.003397 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd" exitCode=0 Nov 25 12:13:24 crc kubenswrapper[4715]: I1125 12:13:24.003472 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae" exitCode=2 Nov 25 12:13:24 crc kubenswrapper[4715]: E1125 12:13:24.217733 4715 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.40:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:24 crc kubenswrapper[4715]: I1125 12:13:24.218880 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:24 crc kubenswrapper[4715]: E1125 12:13:24.254342 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.40:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b3ede77722444 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 12:13:24.253312068 +0000 UTC m=+254.760815089,LastTimestamp:2025-11-25 12:13:24.253312068 +0000 UTC m=+254.760815089,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 12:13:24 crc kubenswrapper[4715]: E1125 12:13:24.653218 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.40:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b3ede77722444 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 12:13:24.253312068 +0000 UTC m=+254.760815089,LastTimestamp:2025-11-25 12:13:24.253312068 +0000 UTC m=+254.760815089,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.017380 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c3f0d6b8992ec3991935c893a99e136f67cebe0ca49014a38c11b8e393b55727"} Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.017723 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"24a9642f301d00a7af08c103b5c54aac691e8be48bdd7ebea5e3825df5a3bcc5"} Nov 25 12:13:25 crc kubenswrapper[4715]: E1125 12:13:25.018482 4715 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.40:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.018519 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.019896 4715 generic.go:334] "Generic (PLEG): container finished" podID="44cf8567-083d-44d6-bab3-a90ee6497454" containerID="2ba43105430a1590dde6d498b904ffeb72b4bb8b9be1e811695d62b3de8e27cd" exitCode=0 Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.019965 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thpk4" event={"ID":"44cf8567-083d-44d6-bab3-a90ee6497454","Type":"ContainerDied","Data":"2ba43105430a1590dde6d498b904ffeb72b4bb8b9be1e811695d62b3de8e27cd"} Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.020505 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.020732 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.022119 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.023157 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.025281 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4" exitCode=0 Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.025320 4715 scope.go:117] "RemoveContainer" containerID="3914e063df931b6092f22f3f729f525511193e9d7624decf00e966a85e8d9d84" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.324423 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.333467 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.333908 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.446613 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kube-api-access\") pod \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.446715 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-var-lock\") pod \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.446812 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-var-lock" (OuterVolumeSpecName: "var-lock") pod "2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" (UID: "2f68897a-2dd5-450c-9bbe-aaf517f6e2b2"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.447036 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kubelet-dir\") pod \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\" (UID: \"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2\") " Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.447117 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" (UID: "2f68897a-2dd5-450c-9bbe-aaf517f6e2b2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.452045 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.452083 4715 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.465447 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" (UID: "2f68897a-2dd5-450c-9bbe-aaf517f6e2b2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:25 crc kubenswrapper[4715]: I1125 12:13:25.553562 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f68897a-2dd5-450c-9bbe-aaf517f6e2b2-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.037304 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjzs4" event={"ID":"b6590884-9e6f-4390-b4c1-59902c2684ac","Type":"ContainerStarted","Data":"35e70c8a6720c7b47bfc07c761c81c0450938af8e0d11b48e6d7bcf074cc9acb"} Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.038520 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f68897a-2dd5-450c-9bbe-aaf517f6e2b2","Type":"ContainerDied","Data":"0c4c299b22b50fb777591366e0065dc9f8748596899ac28bb759fea4b1f2d016"} Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.038544 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c4c299b22b50fb777591366e0065dc9f8748596899ac28bb759fea4b1f2d016" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.038653 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.041539 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.073615 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.074054 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.876911 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.879292 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.879786 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.880128 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.880463 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.972144 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.972244 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.972281 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.972659 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.972697 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:13:26 crc kubenswrapper[4715]: I1125 12:13:26.972717 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.049225 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.050287 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.050436 4715 scope.go:117] "RemoveContainer" containerID="db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.050183 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1" exitCode=0 Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.051205 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.051500 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.051759 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.052002 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.074080 4715 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.074114 4715 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.074126 4715 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.086363 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.086732 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.087051 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.087324 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.209707 4715 scope.go:117] "RemoveContainer" containerID="adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.225100 4715 scope.go:117] "RemoveContainer" containerID="eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.246599 4715 scope.go:117] "RemoveContainer" containerID="97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.267051 4715 scope.go:117] "RemoveContainer" containerID="d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.281680 4715 scope.go:117] "RemoveContainer" containerID="f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.326381 4715 scope.go:117] "RemoveContainer" containerID="db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.326962 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\": container with ID starting with db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4 not found: ID does not exist" containerID="db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.327002 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4"} err="failed to get container status \"db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\": rpc error: code = NotFound desc = could not find container \"db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4\": container with ID starting with db2ee604182950fc4ac1cbada0e45014b9a24c7f531c89758ff514bfd153aca4 not found: ID does not exist" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.327032 4715 scope.go:117] "RemoveContainer" containerID="adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.327454 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\": container with ID starting with adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728 not found: ID does not exist" containerID="adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.327514 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728"} err="failed to get container status \"adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\": rpc error: code = NotFound desc = could not find container \"adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728\": container with ID starting with adbbc9f1f90787f5db2e12b267d4b1f37111d18ad4a2552be1c9357cb7a90728 not found: ID does not exist" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.327552 4715 scope.go:117] "RemoveContainer" containerID="eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.327927 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\": container with ID starting with eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd not found: ID does not exist" containerID="eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.327961 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd"} err="failed to get container status \"eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\": rpc error: code = NotFound desc = could not find container \"eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd\": container with ID starting with eaba646cac7fac1caa5e1adf4ecbd3bfe305a934a0fa6af2c812bf29fd9754fd not found: ID does not exist" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.327981 4715 scope.go:117] "RemoveContainer" containerID="97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.328264 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\": container with ID starting with 97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae not found: ID does not exist" containerID="97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.328318 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae"} err="failed to get container status \"97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\": rpc error: code = NotFound desc = could not find container \"97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae\": container with ID starting with 97fc3251554afb748be6e2792035ba9d7812128f80f9b4aae0836caffea7fbae not found: ID does not exist" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.328340 4715 scope.go:117] "RemoveContainer" containerID="d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.328797 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\": container with ID starting with d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1 not found: ID does not exist" containerID="d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.328866 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1"} err="failed to get container status \"d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\": rpc error: code = NotFound desc = could not find container \"d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1\": container with ID starting with d98d7aa8b96a4c2829311435e32157c9677f14f99c923cb451308c69fbeff8d1 not found: ID does not exist" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.328918 4715 scope.go:117] "RemoveContainer" containerID="f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.329456 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\": container with ID starting with f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c not found: ID does not exist" containerID="f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.329493 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c"} err="failed to get container status \"f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\": rpc error: code = NotFound desc = could not find container \"f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c\": container with ID starting with f5cdff26993b3c540204951191f8b0b1073fa016d6320c023938ad4af935d77c not found: ID does not exist" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.511023 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.511096 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.551148 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.552434 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.553173 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.553561 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.553834 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.554072 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.842518 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.843130 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.843623 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.843882 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.844181 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:27 crc kubenswrapper[4715]: I1125 12:13:27.844235 4715 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 12:13:27 crc kubenswrapper[4715]: E1125 12:13:27.844671 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="200ms" Nov 25 12:13:28 crc kubenswrapper[4715]: E1125 12:13:28.045991 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="400ms" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.059834 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thpk4" event={"ID":"44cf8567-083d-44d6-bab3-a90ee6497454","Type":"ContainerStarted","Data":"321c24746654b1695616a5b4d3657ff0e29530becd1c081691cf20a70880ed60"} Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.060643 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.061002 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.061394 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.061712 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.062031 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.093834 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.093902 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.114008 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zp575" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.114558 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.114879 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.115346 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.115636 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.115844 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.132440 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.132908 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.133339 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.133729 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.133984 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.134254 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.134494 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.355783 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.356307 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.356614 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.356865 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.357099 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.357342 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.357589 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.357823 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:28 crc kubenswrapper[4715]: E1125 12:13:28.367812 4715 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.129.56.40:6443: connect: connection refused" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" volumeName="registry-storage" Nov 25 12:13:28 crc kubenswrapper[4715]: E1125 12:13:28.447146 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="800ms" Nov 25 12:13:28 crc kubenswrapper[4715]: I1125 12:13:28.700756 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.106754 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w95rs" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.107253 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.107497 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.107806 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.108333 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.108543 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.108724 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: E1125 12:13:29.247863 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="1.6s" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.904939 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.905006 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.951431 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.951976 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.952625 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.953282 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.953962 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.954251 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:29 crc kubenswrapper[4715]: I1125 12:13:29.954489 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.109445 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xjzs4" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.110266 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.110802 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.111400 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.111790 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.112067 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.112570 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.528098 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.528151 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.568897 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.569526 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.569994 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.570265 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.570590 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.570981 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.571351 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.696801 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.697071 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.697477 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.698098 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.698396 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: I1125 12:13:30.698678 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:30 crc kubenswrapper[4715]: E1125 12:13:30.849110 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="3.2s" Nov 25 12:13:34 crc kubenswrapper[4715]: E1125 12:13:34.049594 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.40:6443: connect: connection refused" interval="6.4s" Nov 25 12:13:34 crc kubenswrapper[4715]: E1125 12:13:34.655227 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.40:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b3ede77722444 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 12:13:24.253312068 +0000 UTC m=+254.760815089,LastTimestamp:2025-11-25 12:13:24.253312068 +0000 UTC m=+254.760815089,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.692944 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.693642 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.694042 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.695232 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.695762 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.696583 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.697846 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.717444 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.717491 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:34 crc kubenswrapper[4715]: E1125 12:13:34.717963 4715 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:34 crc kubenswrapper[4715]: I1125 12:13:34.718697 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:34 crc kubenswrapper[4715]: W1125 12:13:34.747497 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-85707f3d604321d4d8f4e2cd4543eeb0406d3b4c918d384d61388002bda27a93 WatchSource:0}: Error finding container 85707f3d604321d4d8f4e2cd4543eeb0406d3b4c918d384d61388002bda27a93: Status 404 returned error can't find the container with id 85707f3d604321d4d8f4e2cd4543eeb0406d3b4c918d384d61388002bda27a93 Nov 25 12:13:35 crc kubenswrapper[4715]: I1125 12:13:35.101764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"85707f3d604321d4d8f4e2cd4543eeb0406d3b4c918d384d61388002bda27a93"} Nov 25 12:13:37 crc kubenswrapper[4715]: I1125 12:13:37.112598 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f7dc9dd5eb1ae16f47af3ca14c87bbcbff5db5e84376b1d5b954d18e0e514b13"} Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.122953 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.123219 4715 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637" exitCode=1 Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.123342 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637"} Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.123483 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.123497 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:38 crc kubenswrapper[4715]: E1125 12:13:38.123790 4715 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.123882 4715 scope.go:117] "RemoveContainer" containerID="aae71c3723c1fc90408725a85249a04c8c3b1593664922c1da5f1e9e09289637" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.124514 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.124721 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.125013 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.125500 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.125687 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.125864 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.126026 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.126508 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.126915 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.127112 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.127295 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.127462 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.127624 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:38 crc kubenswrapper[4715]: I1125 12:13:38.127780 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.132852 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.133319 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3662162948ba96348d9a987a62b7cf10a8b8de61ae794846cdb8caf9cb6d17e5"} Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.135822 4715 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="f7dc9dd5eb1ae16f47af3ca14c87bbcbff5db5e84376b1d5b954d18e0e514b13" exitCode=0 Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.135877 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"f7dc9dd5eb1ae16f47af3ca14c87bbcbff5db5e84376b1d5b954d18e0e514b13"} Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.136283 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.136308 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.136781 4715 status_manager.go:851] "Failed to get status for pod" podUID="c4034c63-b17b-4799-a34e-4432e780c343" pod="openshift-marketplace/community-operators-zp575" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zp575\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:39 crc kubenswrapper[4715]: E1125 12:13:39.136962 4715 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.137099 4715 status_manager.go:851] "Failed to get status for pod" podUID="bf866d85-2d8e-4ea2-b1b8-4cd1001beb74" pod="openshift-marketplace/redhat-operators-w95rs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-w95rs\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.137478 4715 status_manager.go:851] "Failed to get status for pod" podUID="44cf8567-083d-44d6-bab3-a90ee6497454" pod="openshift-marketplace/redhat-marketplace-thpk4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-thpk4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.137795 4715 status_manager.go:851] "Failed to get status for pod" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.138406 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.138735 4715 status_manager.go:851] "Failed to get status for pod" podUID="b6590884-9e6f-4390-b4c1-59902c2684ac" pod="openshift-marketplace/certified-operators-xjzs4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-xjzs4\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.139096 4715 status_manager.go:851] "Failed to get status for pod" podUID="df09a836-93f2-4f75-96ab-41b2e0426785" pod="openshift-image-registry/image-registry-66df7c8f76-pc2s5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/image-registry-66df7c8f76-pc2s5\": dial tcp 38.129.56.40:6443: connect: connection refused" Nov 25 12:13:39 crc kubenswrapper[4715]: I1125 12:13:39.653088 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" podUID="09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" containerName="oauth-openshift" containerID="cri-o://822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56" gracePeriod=15 Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.104672 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.144801 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b24884372440228a1458890a6b6ecc7da3b59e0d1e9bf2de2ac3f33a3227128b"} Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.144858 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d03e6a5f771a4f713b10b558bee0387058e853eb394bc8601eb0b136a6617d75"} Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.144868 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1f448176c48945516aa5e51ad023c51eb9bd0cba52d2d04ceaf8de1515f69c20"} Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.146253 4715 generic.go:334] "Generic (PLEG): container finished" podID="09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" containerID="822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56" exitCode=0 Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.146313 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.146321 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" event={"ID":"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31","Type":"ContainerDied","Data":"822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56"} Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.146396 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qztrg" event={"ID":"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31","Type":"ContainerDied","Data":"6049e232f320ad893bb680021fb2cf396b38d34972f10b6a9d44c80278debc0e"} Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.146418 4715 scope.go:117] "RemoveContainer" containerID="822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.174472 4715 scope.go:117] "RemoveContainer" containerID="822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56" Nov 25 12:13:40 crc kubenswrapper[4715]: E1125 12:13:40.177116 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56\": container with ID starting with 822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56 not found: ID does not exist" containerID="822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.177156 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56"} err="failed to get container status \"822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56\": rpc error: code = NotFound desc = could not find container \"822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56\": container with ID starting with 822c950d60ff4386f2fd467933bfa3699efe6e124ae9e81a66da975c8f7f3e56 not found: ID does not exist" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252653 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-policies\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252704 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-router-certs\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252734 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-dir\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252760 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-login\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252794 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-provider-selection\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252818 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-service-ca\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252860 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmkz6\" (UniqueName: \"kubernetes.io/projected/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-kube-api-access-gmkz6\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252889 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-trusted-ca-bundle\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252932 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-cliconfig\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252954 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-serving-cert\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.252986 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-idp-0-file-data\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.253005 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-session\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.253064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-ocp-branding-template\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.253083 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-error\") pod \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\" (UID: \"09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31\") " Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.253897 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.253932 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.254232 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.256281 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.258778 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.327048 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.327805 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.328058 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.329006 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.329371 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-kube-api-access-gmkz6" (OuterVolumeSpecName: "kube-api-access-gmkz6") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "kube-api-access-gmkz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.331565 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.332117 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.332372 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.332545 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" (UID: "09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354765 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354807 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354824 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmkz6\" (UniqueName: \"kubernetes.io/projected/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-kube-api-access-gmkz6\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354838 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354851 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354862 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354874 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354886 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354898 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354909 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354923 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354935 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354948 4715 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.354960 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 12:13:40 crc kubenswrapper[4715]: I1125 12:13:40.565436 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-thpk4" Nov 25 12:13:41 crc kubenswrapper[4715]: I1125 12:13:41.164702 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4420c39e76026223c2ae3192dff6343e8cc22d32fe6a255f65b140156c2866ee"} Nov 25 12:13:41 crc kubenswrapper[4715]: I1125 12:13:41.164758 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2069997d2889c8f871d892ee67a0b48c936450516bb2c16fb06bb2a91fae62b7"} Nov 25 12:13:41 crc kubenswrapper[4715]: I1125 12:13:41.165024 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:41 crc kubenswrapper[4715]: I1125 12:13:41.165043 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:41 crc kubenswrapper[4715]: I1125 12:13:41.165269 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:42 crc kubenswrapper[4715]: I1125 12:13:42.404586 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:13:42 crc kubenswrapper[4715]: I1125 12:13:42.408089 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:13:43 crc kubenswrapper[4715]: I1125 12:13:43.176324 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:13:44 crc kubenswrapper[4715]: I1125 12:13:44.718856 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:44 crc kubenswrapper[4715]: I1125 12:13:44.719154 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:44 crc kubenswrapper[4715]: I1125 12:13:44.724579 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:44 crc kubenswrapper[4715]: I1125 12:13:44.784549 4715 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod55c59d7a-0cdb-48e7-a315-f2a3fdeee4df"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod55c59d7a-0cdb-48e7-a315-f2a3fdeee4df] : Timed out while waiting for systemd to remove kubepods-burstable-pod55c59d7a_0cdb_48e7_a315_f2a3fdeee4df.slice" Nov 25 12:13:46 crc kubenswrapper[4715]: I1125 12:13:46.174978 4715 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:46 crc kubenswrapper[4715]: I1125 12:13:46.243106 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="511ac328-5b87-4b6f-84b6-f4ce0ad67b04" Nov 25 12:13:47 crc kubenswrapper[4715]: I1125 12:13:47.197874 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:47 crc kubenswrapper[4715]: I1125 12:13:47.197913 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:47 crc kubenswrapper[4715]: I1125 12:13:47.201578 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="511ac328-5b87-4b6f-84b6-f4ce0ad67b04" Nov 25 12:13:47 crc kubenswrapper[4715]: I1125 12:13:47.202677 4715 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://1f448176c48945516aa5e51ad023c51eb9bd0cba52d2d04ceaf8de1515f69c20" Nov 25 12:13:47 crc kubenswrapper[4715]: I1125 12:13:47.202715 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:13:48 crc kubenswrapper[4715]: I1125 12:13:48.203546 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:48 crc kubenswrapper[4715]: I1125 12:13:48.203579 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2e09878-9356-4e26-8774-73369b455905" Nov 25 12:13:48 crc kubenswrapper[4715]: I1125 12:13:48.206825 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="511ac328-5b87-4b6f-84b6-f4ce0ad67b04" Nov 25 12:13:53 crc kubenswrapper[4715]: I1125 12:13:53.500744 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 12:13:55 crc kubenswrapper[4715]: I1125 12:13:55.884118 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 12:13:56 crc kubenswrapper[4715]: I1125 12:13:56.962292 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 12:13:56 crc kubenswrapper[4715]: I1125 12:13:56.999260 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 12:13:57 crc kubenswrapper[4715]: I1125 12:13:57.086919 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 12:13:57 crc kubenswrapper[4715]: I1125 12:13:57.224436 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 12:13:57 crc kubenswrapper[4715]: I1125 12:13:57.308013 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 12:13:57 crc kubenswrapper[4715]: I1125 12:13:57.586412 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 12:13:57 crc kubenswrapper[4715]: I1125 12:13:57.775252 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 12:13:57 crc kubenswrapper[4715]: I1125 12:13:57.938592 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.012844 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.052615 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.108103 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.311122 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.407678 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.429372 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.504305 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.537792 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.560722 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.672057 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.706895 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.729293 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.855280 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 12:13:58 crc kubenswrapper[4715]: I1125 12:13:58.943740 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.096257 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.175755 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.187451 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.338839 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.406844 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.587042 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.587041 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.626446 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.674269 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.797831 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.877855 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.886431 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.890510 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.912372 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.925104 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 12:13:59 crc kubenswrapper[4715]: I1125 12:13:59.986099 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.005526 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.018587 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.020023 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.128698 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.220978 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.257246 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.352284 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.365405 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.473648 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.486863 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.503593 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.506964 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.540897 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.706872 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.717917 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.756260 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.807828 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.835091 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.844925 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.878900 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 12:14:00 crc kubenswrapper[4715]: I1125 12:14:00.953337 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.000710 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.005227 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.062107 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.138640 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.243219 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.312800 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.321372 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.330458 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.425830 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.448769 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.456392 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.493237 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.503372 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.509152 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.749402 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.774528 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.793877 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.812721 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.856831 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 12:14:01 crc kubenswrapper[4715]: I1125 12:14:01.860444 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.048794 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.160070 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.169473 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.190246 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.263344 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.303054 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.357626 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.435473 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.440168 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.488811 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.536806 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.554272 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.575090 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.693285 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.836632 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.861804 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.900602 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 12:14:02 crc kubenswrapper[4715]: I1125 12:14:02.935377 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.002623 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.004731 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.108611 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.141267 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.158824 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.178964 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.181800 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.397077 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.413327 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.560561 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.582272 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.649735 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.707313 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.769137 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.879823 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.890308 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.907600 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 12:14:03 crc kubenswrapper[4715]: I1125 12:14:03.953643 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.048339 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.086799 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.136328 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.148384 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.200528 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.246287 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.262761 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.312654 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.327406 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.424701 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.576409 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.629494 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.675985 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.798057 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.929582 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 12:14:04 crc kubenswrapper[4715]: I1125 12:14:04.931376 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.063465 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.095314 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.156511 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.168011 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.234538 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.281557 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.303841 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.306680 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.387550 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.391903 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.484987 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.533655 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.786958 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.899089 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.945882 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 12:14:05 crc kubenswrapper[4715]: I1125 12:14:05.961994 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.147789 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.194848 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.201861 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.217765 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.253411 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.268244 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.284653 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.295809 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.472907 4715 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.476146 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-thpk4" podStartSLOduration=41.374602005 podStartE2EDuration="46.476124761s" podCreationTimestamp="2025-11-25 12:13:20 +0000 UTC" firstStartedPulling="2025-11-25 12:13:22.108706203 +0000 UTC m=+252.616209224" lastFinishedPulling="2025-11-25 12:13:27.210228959 +0000 UTC m=+257.717731980" observedRunningTime="2025-11-25 12:13:46.049927353 +0000 UTC m=+276.557430374" watchObservedRunningTime="2025-11-25 12:14:06.476124761 +0000 UTC m=+296.983627792" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.477112 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xjzs4" podStartSLOduration=43.61578484 podStartE2EDuration="47.477105608s" podCreationTimestamp="2025-11-25 12:13:19 +0000 UTC" firstStartedPulling="2025-11-25 12:13:20.961954373 +0000 UTC m=+251.469457394" lastFinishedPulling="2025-11-25 12:13:24.823275141 +0000 UTC m=+255.330778162" observedRunningTime="2025-11-25 12:13:46.129937583 +0000 UTC m=+276.637440604" watchObservedRunningTime="2025-11-25 12:14:06.477105608 +0000 UTC m=+296.984608629" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.477920 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-qztrg"] Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.477972 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.477991 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kbdhr"] Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.493640 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.493618913 podStartE2EDuration="20.493618913s" podCreationTimestamp="2025-11-25 12:13:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:14:06.492782459 +0000 UTC m=+297.000285490" watchObservedRunningTime="2025-11-25 12:14:06.493618913 +0000 UTC m=+297.001121934" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.651782 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.704795 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" path="/var/lib/kubelet/pods/09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31/volumes" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.729329 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.761284 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.836978 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 12:14:06 crc kubenswrapper[4715]: I1125 12:14:06.872730 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.012754 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.038373 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.067015 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6cb668d466-67mkl"] Nov 25 12:14:07 crc kubenswrapper[4715]: E1125 12:14:07.067300 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" containerName="installer" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.067319 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" containerName="installer" Nov 25 12:14:07 crc kubenswrapper[4715]: E1125 12:14:07.067335 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" containerName="oauth-openshift" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.067343 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" containerName="oauth-openshift" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.067473 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f4dd10-8f2d-4fb8-b3f6-8e72db57ec31" containerName="oauth-openshift" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.067492 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f68897a-2dd5-450c-9bbe-aaf517f6e2b2" containerName="installer" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.067968 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.071334 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.071425 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.071796 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.071909 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.072964 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.073261 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.073283 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.073515 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.076497 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.077852 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.083057 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.083660 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.084052 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.087144 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.089118 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106206 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106316 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106373 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86724\" (UniqueName: \"kubernetes.io/projected/1a9d95e9-cba5-423c-831e-49acd4aa4db4-kube-api-access-86724\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106406 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106430 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-template-login\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106471 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106535 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106609 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-template-error\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106629 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-audit-policies\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106692 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106715 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-session\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106771 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106830 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1a9d95e9-cba5-423c-831e-49acd4aa4db4-audit-dir\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.106890 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.190960 4715 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.208220 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.208272 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86724\" (UniqueName: \"kubernetes.io/projected/1a9d95e9-cba5-423c-831e-49acd4aa4db4-kube-api-access-86724\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.208304 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.208332 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-template-login\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.208375 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.208406 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.208432 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-template-error\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209380 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-audit-policies\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209425 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209436 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209451 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-session\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209472 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209496 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1a9d95e9-cba5-423c-831e-49acd4aa4db4-audit-dir\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209521 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209543 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209782 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1a9d95e9-cba5-423c-831e-49acd4aa4db4-audit-dir\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.209838 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.210046 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-audit-policies\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.210791 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.213801 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.214242 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-template-login\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.215308 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.215683 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-session\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.216941 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.217702 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.223490 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.225451 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1a9d95e9-cba5-423c-831e-49acd4aa4db4-v4-0-config-user-template-error\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.225600 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86724\" (UniqueName: \"kubernetes.io/projected/1a9d95e9-cba5-423c-831e-49acd4aa4db4-kube-api-access-86724\") pod \"oauth-openshift-6cb668d466-67mkl\" (UID: \"1a9d95e9-cba5-423c-831e-49acd4aa4db4\") " pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.338860 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.388970 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.412489 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.486922 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.494472 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.507763 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.581104 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 12:14:07 crc kubenswrapper[4715]: I1125 12:14:07.755751 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.000861 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.088621 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.131675 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.164089 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.170846 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.261151 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.435769 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.472246 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cb668d466-67mkl"] Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.514548 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.681443 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.681577 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.893345 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.918382 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.920585 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.924174 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.931726 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.963111 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 12:14:08 crc kubenswrapper[4715]: I1125 12:14:08.998148 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cb668d466-67mkl"] Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.054868 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.056412 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.152069 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.170167 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.199126 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.237082 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.319128 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" event={"ID":"1a9d95e9-cba5-423c-831e-49acd4aa4db4","Type":"ContainerStarted","Data":"714662b7d47fc81f44afc7f7eec6f8e31893b34bb90f7c50153929cac10c7878"} Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.319796 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" event={"ID":"1a9d95e9-cba5-423c-831e-49acd4aa4db4","Type":"ContainerStarted","Data":"2ba95668520ff1817a258d753b38c41493b2e06059514e6da17a94e3b589463d"} Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.319824 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.341350 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" podStartSLOduration=55.341333953 podStartE2EDuration="55.341333953s" podCreationTimestamp="2025-11-25 12:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:14:09.339879583 +0000 UTC m=+299.847382604" watchObservedRunningTime="2025-11-25 12:14:09.341333953 +0000 UTC m=+299.848836974" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.547575 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.623720 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.629051 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.681802 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.695041 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6cb668d466-67mkl" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.707982 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 12:14:09 crc kubenswrapper[4715]: I1125 12:14:09.852312 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.015241 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.055858 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.073791 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.109332 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.152542 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.273788 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.295381 4715 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.443959 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.536421 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.559011 4715 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.592647 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.629727 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.677661 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.723248 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.755270 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 12:14:10 crc kubenswrapper[4715]: I1125 12:14:10.848852 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.082113 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.108755 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.170517 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.258434 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.294874 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.327622 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.404063 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.489510 4715 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.664065 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.815553 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.886988 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 12:14:11 crc kubenswrapper[4715]: I1125 12:14:11.917114 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 12:14:12 crc kubenswrapper[4715]: I1125 12:14:12.329392 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 12:14:12 crc kubenswrapper[4715]: I1125 12:14:12.338791 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 12:14:12 crc kubenswrapper[4715]: I1125 12:14:12.644767 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 12:14:14 crc kubenswrapper[4715]: I1125 12:14:14.723635 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 12:14:20 crc kubenswrapper[4715]: I1125 12:14:20.023770 4715 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 12:14:20 crc kubenswrapper[4715]: I1125 12:14:20.023996 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://c3f0d6b8992ec3991935c893a99e136f67cebe0ca49014a38c11b8e393b55727" gracePeriod=5 Nov 25 12:14:23 crc kubenswrapper[4715]: I1125 12:14:23.728388 4715 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.401469 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.401532 4715 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="c3f0d6b8992ec3991935c893a99e136f67cebe0ca49014a38c11b8e393b55727" exitCode=137 Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.433383 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.593492 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.593577 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.702597 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.703035 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.702776 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.703103 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.703214 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.703329 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.703391 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.703512 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.703690 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.704021 4715 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.704045 4715 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.704055 4715 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.704077 4715 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.710774 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:14:25 crc kubenswrapper[4715]: I1125 12:14:25.805881 4715 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:26 crc kubenswrapper[4715]: I1125 12:14:26.409558 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 12:14:26 crc kubenswrapper[4715]: I1125 12:14:26.409674 4715 scope.go:117] "RemoveContainer" containerID="c3f0d6b8992ec3991935c893a99e136f67cebe0ca49014a38c11b8e393b55727" Nov 25 12:14:26 crc kubenswrapper[4715]: I1125 12:14:26.409805 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 12:14:26 crc kubenswrapper[4715]: I1125 12:14:26.702675 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.511316 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" podUID="3088de81-03d3-483f-b6c1-ddf0a6e30c34" containerName="registry" containerID="cri-o://d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079" gracePeriod=30 Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.827030 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.897587 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjrnk\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-kube-api-access-sjrnk\") pod \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.897728 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-bound-sa-token\") pod \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.897826 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-certificates\") pod \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.897860 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3088de81-03d3-483f-b6c1-ddf0a6e30c34-ca-trust-extracted\") pod \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.897887 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-tls\") pod \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.897915 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3088de81-03d3-483f-b6c1-ddf0a6e30c34-installation-pull-secrets\") pod \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.897949 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-trusted-ca\") pod \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.898071 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\" (UID: \"3088de81-03d3-483f-b6c1-ddf0a6e30c34\") " Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.898821 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "3088de81-03d3-483f-b6c1-ddf0a6e30c34" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.898871 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "3088de81-03d3-483f-b6c1-ddf0a6e30c34" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.903847 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "3088de81-03d3-483f-b6c1-ddf0a6e30c34" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.904687 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3088de81-03d3-483f-b6c1-ddf0a6e30c34-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "3088de81-03d3-483f-b6c1-ddf0a6e30c34" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.904725 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "3088de81-03d3-483f-b6c1-ddf0a6e30c34" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.905288 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-kube-api-access-sjrnk" (OuterVolumeSpecName: "kube-api-access-sjrnk") pod "3088de81-03d3-483f-b6c1-ddf0a6e30c34" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34"). InnerVolumeSpecName "kube-api-access-sjrnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.914383 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3088de81-03d3-483f-b6c1-ddf0a6e30c34-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "3088de81-03d3-483f-b6c1-ddf0a6e30c34" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.916256 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "3088de81-03d3-483f-b6c1-ddf0a6e30c34" (UID: "3088de81-03d3-483f-b6c1-ddf0a6e30c34"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.999358 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.999411 4715 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.999427 4715 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3088de81-03d3-483f-b6c1-ddf0a6e30c34-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.999440 4715 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.999453 4715 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3088de81-03d3-483f-b6c1-ddf0a6e30c34-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.999465 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3088de81-03d3-483f-b6c1-ddf0a6e30c34-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:31 crc kubenswrapper[4715]: I1125 12:14:31.999477 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjrnk\" (UniqueName: \"kubernetes.io/projected/3088de81-03d3-483f-b6c1-ddf0a6e30c34-kube-api-access-sjrnk\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.441055 4715 generic.go:334] "Generic (PLEG): container finished" podID="3088de81-03d3-483f-b6c1-ddf0a6e30c34" containerID="d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079" exitCode=0 Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.441097 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" event={"ID":"3088de81-03d3-483f-b6c1-ddf0a6e30c34","Type":"ContainerDied","Data":"d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079"} Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.441131 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" event={"ID":"3088de81-03d3-483f-b6c1-ddf0a6e30c34","Type":"ContainerDied","Data":"ab54f30daef82c55ac5e236b047e9ccb17a70f55e0b5c9e87aa739823435f9f3"} Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.441130 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kbdhr" Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.441146 4715 scope.go:117] "RemoveContainer" containerID="d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079" Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.456907 4715 scope.go:117] "RemoveContainer" containerID="d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079" Nov 25 12:14:32 crc kubenswrapper[4715]: E1125 12:14:32.457328 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079\": container with ID starting with d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079 not found: ID does not exist" containerID="d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079" Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.457377 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079"} err="failed to get container status \"d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079\": rpc error: code = NotFound desc = could not find container \"d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079\": container with ID starting with d38a4768a1163184674971c17fef060f2ce67d448fcb75b604f45c5cdf4c6079 not found: ID does not exist" Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.482995 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kbdhr"] Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.489059 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kbdhr"] Nov 25 12:14:32 crc kubenswrapper[4715]: I1125 12:14:32.708935 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3088de81-03d3-483f-b6c1-ddf0a6e30c34" path="/var/lib/kubelet/pods/3088de81-03d3-483f-b6c1-ddf0a6e30c34/volumes" Nov 25 12:14:36 crc kubenswrapper[4715]: I1125 12:14:36.014076 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 12:14:37 crc kubenswrapper[4715]: I1125 12:14:37.481928 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 12:14:38 crc kubenswrapper[4715]: I1125 12:14:38.835086 4715 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.279295 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jtjqf"] Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.279854 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" podUID="4528d803-2051-4abb-9af2-898361f7b2fc" containerName="controller-manager" containerID="cri-o://68050fbf9456c16612dc5eb6d2607e0774dedd8e25cfe7bb2708bfb88d332df2" gracePeriod=30 Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.378151 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck"] Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.378456 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" podUID="4ba7f02a-888d-4def-a53b-7a39407051fb" containerName="route-controller-manager" containerID="cri-o://fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302" gracePeriod=30 Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.503492 4715 generic.go:334] "Generic (PLEG): container finished" podID="4528d803-2051-4abb-9af2-898361f7b2fc" containerID="68050fbf9456c16612dc5eb6d2607e0774dedd8e25cfe7bb2708bfb88d332df2" exitCode=0 Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.503574 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" event={"ID":"4528d803-2051-4abb-9af2-898361f7b2fc","Type":"ContainerDied","Data":"68050fbf9456c16612dc5eb6d2607e0774dedd8e25cfe7bb2708bfb88d332df2"} Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.595594 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.623332 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4528d803-2051-4abb-9af2-898361f7b2fc-serving-cert\") pod \"4528d803-2051-4abb-9af2-898361f7b2fc\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.623384 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-config\") pod \"4528d803-2051-4abb-9af2-898361f7b2fc\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.623418 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-proxy-ca-bundles\") pod \"4528d803-2051-4abb-9af2-898361f7b2fc\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.623447 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-client-ca\") pod \"4528d803-2051-4abb-9af2-898361f7b2fc\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.623570 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvtc6\" (UniqueName: \"kubernetes.io/projected/4528d803-2051-4abb-9af2-898361f7b2fc-kube-api-access-hvtc6\") pod \"4528d803-2051-4abb-9af2-898361f7b2fc\" (UID: \"4528d803-2051-4abb-9af2-898361f7b2fc\") " Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.624490 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-client-ca" (OuterVolumeSpecName: "client-ca") pod "4528d803-2051-4abb-9af2-898361f7b2fc" (UID: "4528d803-2051-4abb-9af2-898361f7b2fc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.624510 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4528d803-2051-4abb-9af2-898361f7b2fc" (UID: "4528d803-2051-4abb-9af2-898361f7b2fc"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.624883 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-config" (OuterVolumeSpecName: "config") pod "4528d803-2051-4abb-9af2-898361f7b2fc" (UID: "4528d803-2051-4abb-9af2-898361f7b2fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.633048 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4528d803-2051-4abb-9af2-898361f7b2fc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4528d803-2051-4abb-9af2-898361f7b2fc" (UID: "4528d803-2051-4abb-9af2-898361f7b2fc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.633767 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4528d803-2051-4abb-9af2-898361f7b2fc-kube-api-access-hvtc6" (OuterVolumeSpecName: "kube-api-access-hvtc6") pod "4528d803-2051-4abb-9af2-898361f7b2fc" (UID: "4528d803-2051-4abb-9af2-898361f7b2fc"). InnerVolumeSpecName "kube-api-access-hvtc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.676697 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.724392 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-config\") pod \"4ba7f02a-888d-4def-a53b-7a39407051fb\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.724446 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-client-ca\") pod \"4ba7f02a-888d-4def-a53b-7a39407051fb\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.725326 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-client-ca" (OuterVolumeSpecName: "client-ca") pod "4ba7f02a-888d-4def-a53b-7a39407051fb" (UID: "4ba7f02a-888d-4def-a53b-7a39407051fb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.725403 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-config" (OuterVolumeSpecName: "config") pod "4ba7f02a-888d-4def-a53b-7a39407051fb" (UID: "4ba7f02a-888d-4def-a53b-7a39407051fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.725481 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba7f02a-888d-4def-a53b-7a39407051fb-serving-cert\") pod \"4ba7f02a-888d-4def-a53b-7a39407051fb\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.725768 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x45kf\" (UniqueName: \"kubernetes.io/projected/4ba7f02a-888d-4def-a53b-7a39407051fb-kube-api-access-x45kf\") pod \"4ba7f02a-888d-4def-a53b-7a39407051fb\" (UID: \"4ba7f02a-888d-4def-a53b-7a39407051fb\") " Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.726423 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvtc6\" (UniqueName: \"kubernetes.io/projected/4528d803-2051-4abb-9af2-898361f7b2fc-kube-api-access-hvtc6\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.726448 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4528d803-2051-4abb-9af2-898361f7b2fc-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.726461 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.726475 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.726486 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4528d803-2051-4abb-9af2-898361f7b2fc-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.726496 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.726505 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ba7f02a-888d-4def-a53b-7a39407051fb-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.729559 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba7f02a-888d-4def-a53b-7a39407051fb-kube-api-access-x45kf" (OuterVolumeSpecName: "kube-api-access-x45kf") pod "4ba7f02a-888d-4def-a53b-7a39407051fb" (UID: "4ba7f02a-888d-4def-a53b-7a39407051fb"). InnerVolumeSpecName "kube-api-access-x45kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.729562 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba7f02a-888d-4def-a53b-7a39407051fb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4ba7f02a-888d-4def-a53b-7a39407051fb" (UID: "4ba7f02a-888d-4def-a53b-7a39407051fb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.827370 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ba7f02a-888d-4def-a53b-7a39407051fb-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:41 crc kubenswrapper[4715]: I1125 12:14:41.827405 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x45kf\" (UniqueName: \"kubernetes.io/projected/4ba7f02a-888d-4def-a53b-7a39407051fb-kube-api-access-x45kf\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.453749 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-544b6b44bb-mmkkl"] Nov 25 12:14:42 crc kubenswrapper[4715]: E1125 12:14:42.454058 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3088de81-03d3-483f-b6c1-ddf0a6e30c34" containerName="registry" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.454075 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3088de81-03d3-483f-b6c1-ddf0a6e30c34" containerName="registry" Nov 25 12:14:42 crc kubenswrapper[4715]: E1125 12:14:42.454099 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4528d803-2051-4abb-9af2-898361f7b2fc" containerName="controller-manager" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.454108 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4528d803-2051-4abb-9af2-898361f7b2fc" containerName="controller-manager" Nov 25 12:14:42 crc kubenswrapper[4715]: E1125 12:14:42.454124 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba7f02a-888d-4def-a53b-7a39407051fb" containerName="route-controller-manager" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.454134 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba7f02a-888d-4def-a53b-7a39407051fb" containerName="route-controller-manager" Nov 25 12:14:42 crc kubenswrapper[4715]: E1125 12:14:42.454154 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.454162 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.454309 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.454321 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba7f02a-888d-4def-a53b-7a39407051fb" containerName="route-controller-manager" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.454333 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4528d803-2051-4abb-9af2-898361f7b2fc" containerName="controller-manager" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.454349 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3088de81-03d3-483f-b6c1-ddf0a6e30c34" containerName="registry" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.454849 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.456736 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5"] Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.457290 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.468466 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-544b6b44bb-mmkkl"] Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.472102 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5"] Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.512294 4715 generic.go:334] "Generic (PLEG): container finished" podID="4ba7f02a-888d-4def-a53b-7a39407051fb" containerID="fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302" exitCode=0 Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.512354 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" event={"ID":"4ba7f02a-888d-4def-a53b-7a39407051fb","Type":"ContainerDied","Data":"fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302"} Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.512370 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.512733 4715 scope.go:117] "RemoveContainer" containerID="fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.512716 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck" event={"ID":"4ba7f02a-888d-4def-a53b-7a39407051fb","Type":"ContainerDied","Data":"2bb9d9b41f11fd3398a587ab3f360da435c0e7950b560570c3715577655f3c56"} Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.517036 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" event={"ID":"4528d803-2051-4abb-9af2-898361f7b2fc","Type":"ContainerDied","Data":"1aab59ec5e23cce40837389d62925f69919a475f2d3d00f3a38be31d8a85547a"} Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.517150 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-jtjqf" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.532402 4715 scope.go:117] "RemoveContainer" containerID="fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302" Nov 25 12:14:42 crc kubenswrapper[4715]: E1125 12:14:42.532864 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302\": container with ID starting with fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302 not found: ID does not exist" containerID="fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.532907 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302"} err="failed to get container status \"fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302\": rpc error: code = NotFound desc = could not find container \"fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302\": container with ID starting with fce51cf90bbca7a6bb154068ee75cfd7c988512429f083ac5c42e2a0d8317302 not found: ID does not exist" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.532941 4715 scope.go:117] "RemoveContainer" containerID="68050fbf9456c16612dc5eb6d2607e0774dedd8e25cfe7bb2708bfb88d332df2" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.536164 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg4r5\" (UniqueName: \"kubernetes.io/projected/6642cf72-138f-4ca2-866e-c583e1c6363b-kube-api-access-qg4r5\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.536256 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57e657db-6cb9-4272-b771-1b86edd37fcd-serving-cert\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.536355 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-config\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.536423 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwch9\" (UniqueName: \"kubernetes.io/projected/57e657db-6cb9-4272-b771-1b86edd37fcd-kube-api-access-rwch9\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.536460 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6642cf72-138f-4ca2-866e-c583e1c6363b-serving-cert\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.536496 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-client-ca\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.536723 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-proxy-ca-bundles\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.536833 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-client-ca\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.536877 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-config\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.538324 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck"] Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.544484 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4kck"] Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.556007 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jtjqf"] Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.559037 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-jtjqf"] Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.638082 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6642cf72-138f-4ca2-866e-c583e1c6363b-serving-cert\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.638267 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-client-ca\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.638336 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-proxy-ca-bundles\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.638375 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-client-ca\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.638416 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-config\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.638481 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg4r5\" (UniqueName: \"kubernetes.io/projected/6642cf72-138f-4ca2-866e-c583e1c6363b-kube-api-access-qg4r5\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.638519 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57e657db-6cb9-4272-b771-1b86edd37fcd-serving-cert\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.638562 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-config\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.638603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwch9\" (UniqueName: \"kubernetes.io/projected/57e657db-6cb9-4272-b771-1b86edd37fcd-kube-api-access-rwch9\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.639608 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-proxy-ca-bundles\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.639639 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-client-ca\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.639678 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-client-ca\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.640135 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-config\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.640647 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-config\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.644517 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6642cf72-138f-4ca2-866e-c583e1c6363b-serving-cert\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.644518 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57e657db-6cb9-4272-b771-1b86edd37fcd-serving-cert\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.659587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg4r5\" (UniqueName: \"kubernetes.io/projected/6642cf72-138f-4ca2-866e-c583e1c6363b-kube-api-access-qg4r5\") pod \"controller-manager-544b6b44bb-mmkkl\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.660008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwch9\" (UniqueName: \"kubernetes.io/projected/57e657db-6cb9-4272-b771-1b86edd37fcd-kube-api-access-rwch9\") pod \"route-controller-manager-7f9f458cd9-nvnt5\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.701722 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4528d803-2051-4abb-9af2-898361f7b2fc" path="/var/lib/kubelet/pods/4528d803-2051-4abb-9af2-898361f7b2fc/volumes" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.702374 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ba7f02a-888d-4def-a53b-7a39407051fb" path="/var/lib/kubelet/pods/4ba7f02a-888d-4def-a53b-7a39407051fb/volumes" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.776782 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.785851 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.966442 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-544b6b44bb-mmkkl"] Nov 25 12:14:42 crc kubenswrapper[4715]: I1125 12:14:42.999142 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5"] Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.309578 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-544b6b44bb-mmkkl"] Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.329988 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5"] Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.523432 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" event={"ID":"57e657db-6cb9-4272-b771-1b86edd37fcd","Type":"ContainerStarted","Data":"9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b"} Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.523480 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" event={"ID":"57e657db-6cb9-4272-b771-1b86edd37fcd","Type":"ContainerStarted","Data":"b245f3a9d36d3647301bc7072910ca100eaceda10faf1c98075d6b0d2be19a34"} Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.528459 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" event={"ID":"6642cf72-138f-4ca2-866e-c583e1c6363b","Type":"ContainerStarted","Data":"6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a"} Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.528504 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" event={"ID":"6642cf72-138f-4ca2-866e-c583e1c6363b","Type":"ContainerStarted","Data":"7d1e1047dfba9ad79de2410a0be2cb5777765ff5e67008e29c2f8905c133e128"} Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.528658 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.534329 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.551429 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" podStartSLOduration=2.551412535 podStartE2EDuration="2.551412535s" podCreationTimestamp="2025-11-25 12:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:14:43.550815858 +0000 UTC m=+334.058318889" watchObservedRunningTime="2025-11-25 12:14:43.551412535 +0000 UTC m=+334.058915556" Nov 25 12:14:43 crc kubenswrapper[4715]: I1125 12:14:43.574359 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" podStartSLOduration=2.574335945 podStartE2EDuration="2.574335945s" podCreationTimestamp="2025-11-25 12:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:14:43.570252728 +0000 UTC m=+334.077755769" watchObservedRunningTime="2025-11-25 12:14:43.574335945 +0000 UTC m=+334.081838966" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.534951 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" podUID="57e657db-6cb9-4272-b771-1b86edd37fcd" containerName="route-controller-manager" containerID="cri-o://9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b" gracePeriod=30 Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.535034 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" podUID="6642cf72-138f-4ca2-866e-c583e1c6363b" containerName="controller-manager" containerID="cri-o://6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a" gracePeriod=30 Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.535097 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.539688 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.927157 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.953333 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn"] Nov 25 12:14:44 crc kubenswrapper[4715]: E1125 12:14:44.953558 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e657db-6cb9-4272-b771-1b86edd37fcd" containerName="route-controller-manager" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.953571 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e657db-6cb9-4272-b771-1b86edd37fcd" containerName="route-controller-manager" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.953671 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e657db-6cb9-4272-b771-1b86edd37fcd" containerName="route-controller-manager" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.954023 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.965885 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn"] Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.973106 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57e657db-6cb9-4272-b771-1b86edd37fcd-serving-cert\") pod \"57e657db-6cb9-4272-b771-1b86edd37fcd\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.973254 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-client-ca\") pod \"57e657db-6cb9-4272-b771-1b86edd37fcd\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.973278 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-config\") pod \"57e657db-6cb9-4272-b771-1b86edd37fcd\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.973299 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwch9\" (UniqueName: \"kubernetes.io/projected/57e657db-6cb9-4272-b771-1b86edd37fcd-kube-api-access-rwch9\") pod \"57e657db-6cb9-4272-b771-1b86edd37fcd\" (UID: \"57e657db-6cb9-4272-b771-1b86edd37fcd\") " Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.975952 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-client-ca" (OuterVolumeSpecName: "client-ca") pod "57e657db-6cb9-4272-b771-1b86edd37fcd" (UID: "57e657db-6cb9-4272-b771-1b86edd37fcd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.981508 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e657db-6cb9-4272-b771-1b86edd37fcd-kube-api-access-rwch9" (OuterVolumeSpecName: "kube-api-access-rwch9") pod "57e657db-6cb9-4272-b771-1b86edd37fcd" (UID: "57e657db-6cb9-4272-b771-1b86edd37fcd"). InnerVolumeSpecName "kube-api-access-rwch9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.981591 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e657db-6cb9-4272-b771-1b86edd37fcd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "57e657db-6cb9-4272-b771-1b86edd37fcd" (UID: "57e657db-6cb9-4272-b771-1b86edd37fcd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:14:44 crc kubenswrapper[4715]: I1125 12:14:44.981911 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-config" (OuterVolumeSpecName: "config") pod "57e657db-6cb9-4272-b771-1b86edd37fcd" (UID: "57e657db-6cb9-4272-b771-1b86edd37fcd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.002982 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074356 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg4r5\" (UniqueName: \"kubernetes.io/projected/6642cf72-138f-4ca2-866e-c583e1c6363b-kube-api-access-qg4r5\") pod \"6642cf72-138f-4ca2-866e-c583e1c6363b\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074443 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6642cf72-138f-4ca2-866e-c583e1c6363b-serving-cert\") pod \"6642cf72-138f-4ca2-866e-c583e1c6363b\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074482 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-proxy-ca-bundles\") pod \"6642cf72-138f-4ca2-866e-c583e1c6363b\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074519 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-client-ca\") pod \"6642cf72-138f-4ca2-866e-c583e1c6363b\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074572 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-config\") pod \"6642cf72-138f-4ca2-866e-c583e1c6363b\" (UID: \"6642cf72-138f-4ca2-866e-c583e1c6363b\") " Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074725 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-client-ca\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074748 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftsg8\" (UniqueName: \"kubernetes.io/projected/4114afad-666b-457a-bc5c-d706839810b1-kube-api-access-ftsg8\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074786 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4114afad-666b-457a-bc5c-d706839810b1-serving-cert\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074809 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-config\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074846 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074857 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57e657db-6cb9-4272-b771-1b86edd37fcd-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074865 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwch9\" (UniqueName: \"kubernetes.io/projected/57e657db-6cb9-4272-b771-1b86edd37fcd-kube-api-access-rwch9\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.074874 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57e657db-6cb9-4272-b771-1b86edd37fcd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.075489 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-config" (OuterVolumeSpecName: "config") pod "6642cf72-138f-4ca2-866e-c583e1c6363b" (UID: "6642cf72-138f-4ca2-866e-c583e1c6363b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.076111 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-client-ca" (OuterVolumeSpecName: "client-ca") pod "6642cf72-138f-4ca2-866e-c583e1c6363b" (UID: "6642cf72-138f-4ca2-866e-c583e1c6363b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.076165 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6642cf72-138f-4ca2-866e-c583e1c6363b" (UID: "6642cf72-138f-4ca2-866e-c583e1c6363b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.077339 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6642cf72-138f-4ca2-866e-c583e1c6363b-kube-api-access-qg4r5" (OuterVolumeSpecName: "kube-api-access-qg4r5") pod "6642cf72-138f-4ca2-866e-c583e1c6363b" (UID: "6642cf72-138f-4ca2-866e-c583e1c6363b"). InnerVolumeSpecName "kube-api-access-qg4r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.078351 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6642cf72-138f-4ca2-866e-c583e1c6363b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6642cf72-138f-4ca2-866e-c583e1c6363b" (UID: "6642cf72-138f-4ca2-866e-c583e1c6363b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.175997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-client-ca\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.176059 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftsg8\" (UniqueName: \"kubernetes.io/projected/4114afad-666b-457a-bc5c-d706839810b1-kube-api-access-ftsg8\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.176124 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4114afad-666b-457a-bc5c-d706839810b1-serving-cert\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.176156 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-config\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.176599 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.176634 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.176647 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6642cf72-138f-4ca2-866e-c583e1c6363b-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.176657 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg4r5\" (UniqueName: \"kubernetes.io/projected/6642cf72-138f-4ca2-866e-c583e1c6363b-kube-api-access-qg4r5\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.176668 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6642cf72-138f-4ca2-866e-c583e1c6363b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.177632 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-client-ca\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.177830 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-config\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.182863 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4114afad-666b-457a-bc5c-d706839810b1-serving-cert\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.198898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftsg8\" (UniqueName: \"kubernetes.io/projected/4114afad-666b-457a-bc5c-d706839810b1-kube-api-access-ftsg8\") pod \"route-controller-manager-6976bb8b7b-67cbn\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.316732 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.481216 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn"] Nov 25 12:14:45 crc kubenswrapper[4715]: W1125 12:14:45.482864 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4114afad_666b_457a_bc5c_d706839810b1.slice/crio-81ce2bb1deaa55fddc93ff24ed21ca78d7c08582062065389a2bc95b2494a47a WatchSource:0}: Error finding container 81ce2bb1deaa55fddc93ff24ed21ca78d7c08582062065389a2bc95b2494a47a: Status 404 returned error can't find the container with id 81ce2bb1deaa55fddc93ff24ed21ca78d7c08582062065389a2bc95b2494a47a Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.540886 4715 generic.go:334] "Generic (PLEG): container finished" podID="57e657db-6cb9-4272-b771-1b86edd37fcd" containerID="9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b" exitCode=0 Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.540935 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" event={"ID":"57e657db-6cb9-4272-b771-1b86edd37fcd","Type":"ContainerDied","Data":"9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b"} Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.540959 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.540996 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5" event={"ID":"57e657db-6cb9-4272-b771-1b86edd37fcd","Type":"ContainerDied","Data":"b245f3a9d36d3647301bc7072910ca100eaceda10faf1c98075d6b0d2be19a34"} Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.541017 4715 scope.go:117] "RemoveContainer" containerID="9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.543948 4715 generic.go:334] "Generic (PLEG): container finished" podID="6642cf72-138f-4ca2-866e-c583e1c6363b" containerID="6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a" exitCode=0 Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.543978 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.544020 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" event={"ID":"6642cf72-138f-4ca2-866e-c583e1c6363b","Type":"ContainerDied","Data":"6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a"} Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.544056 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-544b6b44bb-mmkkl" event={"ID":"6642cf72-138f-4ca2-866e-c583e1c6363b","Type":"ContainerDied","Data":"7d1e1047dfba9ad79de2410a0be2cb5777765ff5e67008e29c2f8905c133e128"} Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.545650 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" event={"ID":"4114afad-666b-457a-bc5c-d706839810b1","Type":"ContainerStarted","Data":"81ce2bb1deaa55fddc93ff24ed21ca78d7c08582062065389a2bc95b2494a47a"} Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.567437 4715 scope.go:117] "RemoveContainer" containerID="9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b" Nov 25 12:14:45 crc kubenswrapper[4715]: E1125 12:14:45.568488 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b\": container with ID starting with 9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b not found: ID does not exist" containerID="9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.568553 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b"} err="failed to get container status \"9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b\": rpc error: code = NotFound desc = could not find container \"9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b\": container with ID starting with 9a482f1502fd659191eaf41a83f2fbea0754dc1543488a8ae243ae7e2fd0571b not found: ID does not exist" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.568590 4715 scope.go:117] "RemoveContainer" containerID="6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.575473 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5"] Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.590620 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f9f458cd9-nvnt5"] Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.597788 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-544b6b44bb-mmkkl"] Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.598802 4715 scope.go:117] "RemoveContainer" containerID="6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a" Nov 25 12:14:45 crc kubenswrapper[4715]: E1125 12:14:45.600024 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a\": container with ID starting with 6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a not found: ID does not exist" containerID="6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.600175 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a"} err="failed to get container status \"6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a\": rpc error: code = NotFound desc = could not find container \"6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a\": container with ID starting with 6931d904d04d88772fc8f1ce481966ac86a9bf47e9c9c287ed601212630d3f9a not found: ID does not exist" Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.606596 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-544b6b44bb-mmkkl"] Nov 25 12:14:45 crc kubenswrapper[4715]: I1125 12:14:45.843035 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 12:14:46 crc kubenswrapper[4715]: I1125 12:14:46.210821 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 12:14:46 crc kubenswrapper[4715]: I1125 12:14:46.554725 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" event={"ID":"4114afad-666b-457a-bc5c-d706839810b1","Type":"ContainerStarted","Data":"b9b73b45c4513a7f65c7d3cd84fcfd064d224af8fff12450cd4b712df8b9c3f8"} Nov 25 12:14:46 crc kubenswrapper[4715]: I1125 12:14:46.556142 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:46 crc kubenswrapper[4715]: I1125 12:14:46.563171 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:14:46 crc kubenswrapper[4715]: I1125 12:14:46.578051 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" podStartSLOduration=3.578013704 podStartE2EDuration="3.578013704s" podCreationTimestamp="2025-11-25 12:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:14:46.573169084 +0000 UTC m=+337.080672105" watchObservedRunningTime="2025-11-25 12:14:46.578013704 +0000 UTC m=+337.085516775" Nov 25 12:14:46 crc kubenswrapper[4715]: I1125 12:14:46.702385 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57e657db-6cb9-4272-b771-1b86edd37fcd" path="/var/lib/kubelet/pods/57e657db-6cb9-4272-b771-1b86edd37fcd/volumes" Nov 25 12:14:46 crc kubenswrapper[4715]: I1125 12:14:46.702914 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6642cf72-138f-4ca2-866e-c583e1c6363b" path="/var/lib/kubelet/pods/6642cf72-138f-4ca2-866e-c583e1c6363b/volumes" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.460995 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b7c486df5-mzqrx"] Nov 25 12:14:47 crc kubenswrapper[4715]: E1125 12:14:47.461344 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6642cf72-138f-4ca2-866e-c583e1c6363b" containerName="controller-manager" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.461366 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6642cf72-138f-4ca2-866e-c583e1c6363b" containerName="controller-manager" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.461547 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6642cf72-138f-4ca2-866e-c583e1c6363b" containerName="controller-manager" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.462093 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.465329 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.465930 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.467429 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.468107 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.468595 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.471110 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.476079 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b7c486df5-mzqrx"] Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.479277 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.521704 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-client-ca\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.521826 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-proxy-ca-bundles\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.521858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zndkv\" (UniqueName: \"kubernetes.io/projected/87722bc8-f42e-43bd-a86a-99ce8fdf4549-kube-api-access-zndkv\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.521936 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-config\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.521960 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87722bc8-f42e-43bd-a86a-99ce8fdf4549-serving-cert\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.623744 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-client-ca\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.623897 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-proxy-ca-bundles\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.623934 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zndkv\" (UniqueName: \"kubernetes.io/projected/87722bc8-f42e-43bd-a86a-99ce8fdf4549-kube-api-access-zndkv\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.623985 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-config\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.624025 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87722bc8-f42e-43bd-a86a-99ce8fdf4549-serving-cert\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.624950 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-client-ca\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.625520 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-proxy-ca-bundles\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.625533 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-config\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.637265 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87722bc8-f42e-43bd-a86a-99ce8fdf4549-serving-cert\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.646594 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zndkv\" (UniqueName: \"kubernetes.io/projected/87722bc8-f42e-43bd-a86a-99ce8fdf4549-kube-api-access-zndkv\") pod \"controller-manager-b7c486df5-mzqrx\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:47 crc kubenswrapper[4715]: I1125 12:14:47.785359 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:48 crc kubenswrapper[4715]: I1125 12:14:48.046634 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b7c486df5-mzqrx"] Nov 25 12:14:48 crc kubenswrapper[4715]: I1125 12:14:48.570579 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" event={"ID":"87722bc8-f42e-43bd-a86a-99ce8fdf4549","Type":"ContainerStarted","Data":"e4d4714ce819ab7b63df92a608d7f6aedd4a859eceaf7028668aba1fac5b5885"} Nov 25 12:14:48 crc kubenswrapper[4715]: I1125 12:14:48.570886 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" event={"ID":"87722bc8-f42e-43bd-a86a-99ce8fdf4549","Type":"ContainerStarted","Data":"e4a1782b00cdb0ea3c4de324213be847800487924c7200b43f1ae9b162ad89cc"} Nov 25 12:14:48 crc kubenswrapper[4715]: I1125 12:14:48.571124 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:48 crc kubenswrapper[4715]: I1125 12:14:48.575317 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:14:48 crc kubenswrapper[4715]: I1125 12:14:48.589110 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" podStartSLOduration=5.589091795 podStartE2EDuration="5.589091795s" podCreationTimestamp="2025-11-25 12:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:14:48.583309738 +0000 UTC m=+339.090812759" watchObservedRunningTime="2025-11-25 12:14:48.589091795 +0000 UTC m=+339.096594826" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.162595 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv"] Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.163881 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.168902 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.168984 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.177092 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv"] Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.279698 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqbds\" (UniqueName: \"kubernetes.io/projected/2b215c8d-689d-431d-815b-b57707fb07ba-kube-api-access-jqbds\") pod \"collect-profiles-29401215-r9rkv\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.280012 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b215c8d-689d-431d-815b-b57707fb07ba-config-volume\") pod \"collect-profiles-29401215-r9rkv\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.280035 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b215c8d-689d-431d-815b-b57707fb07ba-secret-volume\") pod \"collect-profiles-29401215-r9rkv\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.380912 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqbds\" (UniqueName: \"kubernetes.io/projected/2b215c8d-689d-431d-815b-b57707fb07ba-kube-api-access-jqbds\") pod \"collect-profiles-29401215-r9rkv\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.380992 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b215c8d-689d-431d-815b-b57707fb07ba-config-volume\") pod \"collect-profiles-29401215-r9rkv\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.381011 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b215c8d-689d-431d-815b-b57707fb07ba-secret-volume\") pod \"collect-profiles-29401215-r9rkv\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.381992 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b215c8d-689d-431d-815b-b57707fb07ba-config-volume\") pod \"collect-profiles-29401215-r9rkv\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.389675 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b215c8d-689d-431d-815b-b57707fb07ba-secret-volume\") pod \"collect-profiles-29401215-r9rkv\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.398538 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqbds\" (UniqueName: \"kubernetes.io/projected/2b215c8d-689d-431d-815b-b57707fb07ba-kube-api-access-jqbds\") pod \"collect-profiles-29401215-r9rkv\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.492733 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:00 crc kubenswrapper[4715]: I1125 12:15:00.871144 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv"] Nov 25 12:15:01 crc kubenswrapper[4715]: I1125 12:15:01.274285 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b7c486df5-mzqrx"] Nov 25 12:15:01 crc kubenswrapper[4715]: I1125 12:15:01.274785 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" podUID="87722bc8-f42e-43bd-a86a-99ce8fdf4549" containerName="controller-manager" containerID="cri-o://e4d4714ce819ab7b63df92a608d7f6aedd4a859eceaf7028668aba1fac5b5885" gracePeriod=30 Nov 25 12:15:01 crc kubenswrapper[4715]: I1125 12:15:01.282503 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn"] Nov 25 12:15:01 crc kubenswrapper[4715]: I1125 12:15:01.282937 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" podUID="4114afad-666b-457a-bc5c-d706839810b1" containerName="route-controller-manager" containerID="cri-o://b9b73b45c4513a7f65c7d3cd84fcfd064d224af8fff12450cd4b712df8b9c3f8" gracePeriod=30 Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.637733 4715 generic.go:334] "Generic (PLEG): container finished" podID="4114afad-666b-457a-bc5c-d706839810b1" containerID="b9b73b45c4513a7f65c7d3cd84fcfd064d224af8fff12450cd4b712df8b9c3f8" exitCode=0 Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.637829 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" event={"ID":"4114afad-666b-457a-bc5c-d706839810b1","Type":"ContainerDied","Data":"b9b73b45c4513a7f65c7d3cd84fcfd064d224af8fff12450cd4b712df8b9c3f8"} Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.639702 4715 generic.go:334] "Generic (PLEG): container finished" podID="2b215c8d-689d-431d-815b-b57707fb07ba" containerID="fafff72d5e533faab596adc9e437570d59186e0bd71954fbd5e8e495a8d79851" exitCode=0 Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.639807 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" event={"ID":"2b215c8d-689d-431d-815b-b57707fb07ba","Type":"ContainerDied","Data":"fafff72d5e533faab596adc9e437570d59186e0bd71954fbd5e8e495a8d79851"} Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.639835 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" event={"ID":"2b215c8d-689d-431d-815b-b57707fb07ba","Type":"ContainerStarted","Data":"98f2bfb4259954cac4c499553206debe79a98c450a3de2998cb8058969366d00"} Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.641453 4715 generic.go:334] "Generic (PLEG): container finished" podID="87722bc8-f42e-43bd-a86a-99ce8fdf4549" containerID="e4d4714ce819ab7b63df92a608d7f6aedd4a859eceaf7028668aba1fac5b5885" exitCode=0 Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.641507 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" event={"ID":"87722bc8-f42e-43bd-a86a-99ce8fdf4549","Type":"ContainerDied","Data":"e4d4714ce819ab7b63df92a608d7f6aedd4a859eceaf7028668aba1fac5b5885"} Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.835734 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.842793 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.897621 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-proxy-ca-bundles\") pod \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.897710 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-config\") pod \"4114afad-666b-457a-bc5c-d706839810b1\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.897739 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4114afad-666b-457a-bc5c-d706839810b1-serving-cert\") pod \"4114afad-666b-457a-bc5c-d706839810b1\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.897762 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftsg8\" (UniqueName: \"kubernetes.io/projected/4114afad-666b-457a-bc5c-d706839810b1-kube-api-access-ftsg8\") pod \"4114afad-666b-457a-bc5c-d706839810b1\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.897784 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-client-ca\") pod \"4114afad-666b-457a-bc5c-d706839810b1\" (UID: \"4114afad-666b-457a-bc5c-d706839810b1\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.897836 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zndkv\" (UniqueName: \"kubernetes.io/projected/87722bc8-f42e-43bd-a86a-99ce8fdf4549-kube-api-access-zndkv\") pod \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.897868 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-client-ca\") pod \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.897946 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87722bc8-f42e-43bd-a86a-99ce8fdf4549-serving-cert\") pod \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.897971 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-config\") pod \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\" (UID: \"87722bc8-f42e-43bd-a86a-99ce8fdf4549\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.898805 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-client-ca" (OuterVolumeSpecName: "client-ca") pod "4114afad-666b-457a-bc5c-d706839810b1" (UID: "4114afad-666b-457a-bc5c-d706839810b1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.898822 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "87722bc8-f42e-43bd-a86a-99ce8fdf4549" (UID: "87722bc8-f42e-43bd-a86a-99ce8fdf4549"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.898854 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-config" (OuterVolumeSpecName: "config") pod "4114afad-666b-457a-bc5c-d706839810b1" (UID: "4114afad-666b-457a-bc5c-d706839810b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.899088 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-client-ca" (OuterVolumeSpecName: "client-ca") pod "87722bc8-f42e-43bd-a86a-99ce8fdf4549" (UID: "87722bc8-f42e-43bd-a86a-99ce8fdf4549"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.899469 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-config" (OuterVolumeSpecName: "config") pod "87722bc8-f42e-43bd-a86a-99ce8fdf4549" (UID: "87722bc8-f42e-43bd-a86a-99ce8fdf4549"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.903744 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87722bc8-f42e-43bd-a86a-99ce8fdf4549-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "87722bc8-f42e-43bd-a86a-99ce8fdf4549" (UID: "87722bc8-f42e-43bd-a86a-99ce8fdf4549"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.903795 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4114afad-666b-457a-bc5c-d706839810b1-kube-api-access-ftsg8" (OuterVolumeSpecName: "kube-api-access-ftsg8") pod "4114afad-666b-457a-bc5c-d706839810b1" (UID: "4114afad-666b-457a-bc5c-d706839810b1"). InnerVolumeSpecName "kube-api-access-ftsg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.903919 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4114afad-666b-457a-bc5c-d706839810b1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4114afad-666b-457a-bc5c-d706839810b1" (UID: "4114afad-666b-457a-bc5c-d706839810b1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.903992 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87722bc8-f42e-43bd-a86a-99ce8fdf4549-kube-api-access-zndkv" (OuterVolumeSpecName: "kube-api-access-zndkv") pod "87722bc8-f42e-43bd-a86a-99ce8fdf4549" (UID: "87722bc8-f42e-43bd-a86a-99ce8fdf4549"). InnerVolumeSpecName "kube-api-access-zndkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.999028 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.999062 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4114afad-666b-457a-bc5c-d706839810b1-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.999077 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4114afad-666b-457a-bc5c-d706839810b1-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.999090 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftsg8\" (UniqueName: \"kubernetes.io/projected/4114afad-666b-457a-bc5c-d706839810b1-kube-api-access-ftsg8\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.999105 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zndkv\" (UniqueName: \"kubernetes.io/projected/87722bc8-f42e-43bd-a86a-99ce8fdf4549-kube-api-access-zndkv\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.999117 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.999128 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87722bc8-f42e-43bd-a86a-99ce8fdf4549-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.999141 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:01.999152 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87722bc8-f42e-43bd-a86a-99ce8fdf4549-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.464088 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff"] Nov 25 12:15:02 crc kubenswrapper[4715]: E1125 12:15:02.464584 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4114afad-666b-457a-bc5c-d706839810b1" containerName="route-controller-manager" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.464595 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4114afad-666b-457a-bc5c-d706839810b1" containerName="route-controller-manager" Nov 25 12:15:02 crc kubenswrapper[4715]: E1125 12:15:02.464606 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87722bc8-f42e-43bd-a86a-99ce8fdf4549" containerName="controller-manager" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.464614 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="87722bc8-f42e-43bd-a86a-99ce8fdf4549" containerName="controller-manager" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.464714 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="87722bc8-f42e-43bd-a86a-99ce8fdf4549" containerName="controller-manager" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.464726 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4114afad-666b-457a-bc5c-d706839810b1" containerName="route-controller-manager" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.465130 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.468002 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq"] Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.468712 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.475543 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff"] Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.487865 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq"] Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.505978 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab237a51-6c25-4821-91ae-424ab9e437f6-config\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.506017 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a11743f-6a49-4680-8816-38d44e090071-client-ca\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.506059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a11743f-6a49-4680-8816-38d44e090071-serving-cert\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.506118 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab237a51-6c25-4821-91ae-424ab9e437f6-serving-cert\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.506154 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a11743f-6a49-4680-8816-38d44e090071-config\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.506209 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab237a51-6c25-4821-91ae-424ab9e437f6-client-ca\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.506240 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx76g\" (UniqueName: \"kubernetes.io/projected/ab237a51-6c25-4821-91ae-424ab9e437f6-kube-api-access-cx76g\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.506267 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wq58\" (UniqueName: \"kubernetes.io/projected/9a11743f-6a49-4680-8816-38d44e090071-kube-api-access-9wq58\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.506297 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab237a51-6c25-4821-91ae-424ab9e437f6-proxy-ca-bundles\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.606990 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab237a51-6c25-4821-91ae-424ab9e437f6-proxy-ca-bundles\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.607067 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab237a51-6c25-4821-91ae-424ab9e437f6-config\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.607090 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a11743f-6a49-4680-8816-38d44e090071-client-ca\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.607133 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a11743f-6a49-4680-8816-38d44e090071-serving-cert\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.607158 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab237a51-6c25-4821-91ae-424ab9e437f6-serving-cert\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.607232 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a11743f-6a49-4680-8816-38d44e090071-config\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.607317 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab237a51-6c25-4821-91ae-424ab9e437f6-client-ca\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.607352 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx76g\" (UniqueName: \"kubernetes.io/projected/ab237a51-6c25-4821-91ae-424ab9e437f6-kube-api-access-cx76g\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.607374 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wq58\" (UniqueName: \"kubernetes.io/projected/9a11743f-6a49-4680-8816-38d44e090071-kube-api-access-9wq58\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.608278 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9a11743f-6a49-4680-8816-38d44e090071-client-ca\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.608369 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab237a51-6c25-4821-91ae-424ab9e437f6-proxy-ca-bundles\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.608697 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab237a51-6c25-4821-91ae-424ab9e437f6-client-ca\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.608711 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a11743f-6a49-4680-8816-38d44e090071-config\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.610725 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab237a51-6c25-4821-91ae-424ab9e437f6-config\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.610851 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a11743f-6a49-4680-8816-38d44e090071-serving-cert\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.611353 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab237a51-6c25-4821-91ae-424ab9e437f6-serving-cert\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.621427 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wq58\" (UniqueName: \"kubernetes.io/projected/9a11743f-6a49-4680-8816-38d44e090071-kube-api-access-9wq58\") pod \"route-controller-manager-59b77b548f-wx9kq\" (UID: \"9a11743f-6a49-4680-8816-38d44e090071\") " pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.621927 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx76g\" (UniqueName: \"kubernetes.io/projected/ab237a51-6c25-4821-91ae-424ab9e437f6-kube-api-access-cx76g\") pod \"controller-manager-5dfcd8f55d-ktcff\" (UID: \"ab237a51-6c25-4821-91ae-424ab9e437f6\") " pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.648624 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.648615 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b7c486df5-mzqrx" event={"ID":"87722bc8-f42e-43bd-a86a-99ce8fdf4549","Type":"ContainerDied","Data":"e4a1782b00cdb0ea3c4de324213be847800487924c7200b43f1ae9b162ad89cc"} Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.648701 4715 scope.go:117] "RemoveContainer" containerID="e4d4714ce819ab7b63df92a608d7f6aedd4a859eceaf7028668aba1fac5b5885" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.651375 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.651505 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn" event={"ID":"4114afad-666b-457a-bc5c-d706839810b1","Type":"ContainerDied","Data":"81ce2bb1deaa55fddc93ff24ed21ca78d7c08582062065389a2bc95b2494a47a"} Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.664958 4715 scope.go:117] "RemoveContainer" containerID="b9b73b45c4513a7f65c7d3cd84fcfd064d224af8fff12450cd4b712df8b9c3f8" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.682583 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b7c486df5-mzqrx"] Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.689114 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-b7c486df5-mzqrx"] Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.710542 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87722bc8-f42e-43bd-a86a-99ce8fdf4549" path="/var/lib/kubelet/pods/87722bc8-f42e-43bd-a86a-99ce8fdf4549/volumes" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.710967 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn"] Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.710994 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6976bb8b7b-67cbn"] Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.789591 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.800231 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.848037 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.909829 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b215c8d-689d-431d-815b-b57707fb07ba-config-volume\") pod \"2b215c8d-689d-431d-815b-b57707fb07ba\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.910169 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b215c8d-689d-431d-815b-b57707fb07ba-secret-volume\") pod \"2b215c8d-689d-431d-815b-b57707fb07ba\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.910223 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqbds\" (UniqueName: \"kubernetes.io/projected/2b215c8d-689d-431d-815b-b57707fb07ba-kube-api-access-jqbds\") pod \"2b215c8d-689d-431d-815b-b57707fb07ba\" (UID: \"2b215c8d-689d-431d-815b-b57707fb07ba\") " Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.910739 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b215c8d-689d-431d-815b-b57707fb07ba-config-volume" (OuterVolumeSpecName: "config-volume") pod "2b215c8d-689d-431d-815b-b57707fb07ba" (UID: "2b215c8d-689d-431d-815b-b57707fb07ba"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.915315 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b215c8d-689d-431d-815b-b57707fb07ba-kube-api-access-jqbds" (OuterVolumeSpecName: "kube-api-access-jqbds") pod "2b215c8d-689d-431d-815b-b57707fb07ba" (UID: "2b215c8d-689d-431d-815b-b57707fb07ba"). InnerVolumeSpecName "kube-api-access-jqbds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.918961 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b215c8d-689d-431d-815b-b57707fb07ba-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2b215c8d-689d-431d-815b-b57707fb07ba" (UID: "2b215c8d-689d-431d-815b-b57707fb07ba"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:15:02 crc kubenswrapper[4715]: I1125 12:15:02.961154 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff"] Nov 25 12:15:02 crc kubenswrapper[4715]: W1125 12:15:02.972528 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab237a51_6c25_4821_91ae_424ab9e437f6.slice/crio-791c7c86624726d8f8144fa803fcacc0ac08a0db629e94eaadadbd888ff426f0 WatchSource:0}: Error finding container 791c7c86624726d8f8144fa803fcacc0ac08a0db629e94eaadadbd888ff426f0: Status 404 returned error can't find the container with id 791c7c86624726d8f8144fa803fcacc0ac08a0db629e94eaadadbd888ff426f0 Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.011515 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2b215c8d-689d-431d-815b-b57707fb07ba-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.011550 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2b215c8d-689d-431d-815b-b57707fb07ba-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.011563 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqbds\" (UniqueName: \"kubernetes.io/projected/2b215c8d-689d-431d-815b-b57707fb07ba-kube-api-access-jqbds\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.024007 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq"] Nov 25 12:15:03 crc kubenswrapper[4715]: W1125 12:15:03.037743 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a11743f_6a49_4680_8816_38d44e090071.slice/crio-95c706db7ed4a84fd26dc30587be282aeeaf837f2ab2b94eff3daf2af1046880 WatchSource:0}: Error finding container 95c706db7ed4a84fd26dc30587be282aeeaf837f2ab2b94eff3daf2af1046880: Status 404 returned error can't find the container with id 95c706db7ed4a84fd26dc30587be282aeeaf837f2ab2b94eff3daf2af1046880 Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.662265 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.662321 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-r9rkv" event={"ID":"2b215c8d-689d-431d-815b-b57707fb07ba","Type":"ContainerDied","Data":"98f2bfb4259954cac4c499553206debe79a98c450a3de2998cb8058969366d00"} Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.662840 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98f2bfb4259954cac4c499553206debe79a98c450a3de2998cb8058969366d00" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.664032 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" event={"ID":"9a11743f-6a49-4680-8816-38d44e090071","Type":"ContainerStarted","Data":"b8a438dd9d2b3bb7e5ee8df75f8ccad388f9a247c150098c7c8d0433dbae0feb"} Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.664078 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" event={"ID":"9a11743f-6a49-4680-8816-38d44e090071","Type":"ContainerStarted","Data":"95c706db7ed4a84fd26dc30587be282aeeaf837f2ab2b94eff3daf2af1046880"} Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.665402 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.671819 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.674346 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" event={"ID":"ab237a51-6c25-4821-91ae-424ab9e437f6","Type":"ContainerStarted","Data":"043ae51a0c2dd97f4c1f7cf4219e229e450f7ca79d3ff62c8cceafe700aadcc4"} Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.674391 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" event={"ID":"ab237a51-6c25-4821-91ae-424ab9e437f6","Type":"ContainerStarted","Data":"791c7c86624726d8f8144fa803fcacc0ac08a0db629e94eaadadbd888ff426f0"} Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.674712 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.685080 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.692492 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59b77b548f-wx9kq" podStartSLOduration=2.692468864 podStartE2EDuration="2.692468864s" podCreationTimestamp="2025-11-25 12:15:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:15:03.680371005 +0000 UTC m=+354.187874026" watchObservedRunningTime="2025-11-25 12:15:03.692468864 +0000 UTC m=+354.199971885" Nov 25 12:15:03 crc kubenswrapper[4715]: I1125 12:15:03.720202 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5dfcd8f55d-ktcff" podStartSLOduration=2.720166282 podStartE2EDuration="2.720166282s" podCreationTimestamp="2025-11-25 12:15:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:15:03.71975039 +0000 UTC m=+354.227253411" watchObservedRunningTime="2025-11-25 12:15:03.720166282 +0000 UTC m=+354.227669293" Nov 25 12:15:04 crc kubenswrapper[4715]: I1125 12:15:04.699717 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4114afad-666b-457a-bc5c-d706839810b1" path="/var/lib/kubelet/pods/4114afad-666b-457a-bc5c-d706839810b1/volumes" Nov 25 12:15:17 crc kubenswrapper[4715]: I1125 12:15:17.943494 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:15:17 crc kubenswrapper[4715]: I1125 12:15:17.944111 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:15:47 crc kubenswrapper[4715]: I1125 12:15:47.944508 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:15:47 crc kubenswrapper[4715]: I1125 12:15:47.945050 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:16:17 crc kubenswrapper[4715]: I1125 12:16:17.943626 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:16:17 crc kubenswrapper[4715]: I1125 12:16:17.944268 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:16:17 crc kubenswrapper[4715]: I1125 12:16:17.944326 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:16:17 crc kubenswrapper[4715]: I1125 12:16:17.944960 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"133f6a3b52123e9fe263e667ae6646c6fa53439cce70b559acdf67b4d86c9ee6"} pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:16:17 crc kubenswrapper[4715]: I1125 12:16:17.945017 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" containerID="cri-o://133f6a3b52123e9fe263e667ae6646c6fa53439cce70b559acdf67b4d86c9ee6" gracePeriod=600 Nov 25 12:16:18 crc kubenswrapper[4715]: I1125 12:16:18.098623 4715 generic.go:334] "Generic (PLEG): container finished" podID="51676837-adc4-4424-b527-920a6528b6a2" containerID="133f6a3b52123e9fe263e667ae6646c6fa53439cce70b559acdf67b4d86c9ee6" exitCode=0 Nov 25 12:16:18 crc kubenswrapper[4715]: I1125 12:16:18.098917 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerDied","Data":"133f6a3b52123e9fe263e667ae6646c6fa53439cce70b559acdf67b4d86c9ee6"} Nov 25 12:16:18 crc kubenswrapper[4715]: I1125 12:16:18.099112 4715 scope.go:117] "RemoveContainer" containerID="2e0a997c36f959cab42012ad6adaf2a07c7067aefb3743c0cd1b6ecc667a5e71" Nov 25 12:16:19 crc kubenswrapper[4715]: I1125 12:16:19.109732 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"d16257d5ac23a2866262558b39c2b844ea7649b0c06fa81698cc8fe1fd3c8f95"} Nov 25 12:18:47 crc kubenswrapper[4715]: I1125 12:18:47.943647 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:18:47 crc kubenswrapper[4715]: I1125 12:18:47.944458 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.382657 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tzl95"] Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.383661 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovn-controller" containerID="cri-o://2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3" gracePeriod=30 Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.383726 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="nbdb" containerID="cri-o://d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e" gracePeriod=30 Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.383792 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kube-rbac-proxy-node" containerID="cri-o://8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31" gracePeriod=30 Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.383874 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772" gracePeriod=30 Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.383851 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="northd" containerID="cri-o://6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e" gracePeriod=30 Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.383843 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="sbdb" containerID="cri-o://543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15" gracePeriod=30 Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.383795 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovn-acl-logging" containerID="cri-o://3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149" gracePeriod=30 Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.439764 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" containerID="cri-o://40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e" gracePeriod=30 Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.470457 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/2.log" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.471518 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/1.log" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.471573 4715 generic.go:334] "Generic (PLEG): container finished" podID="6371ac18-2361-43bb-b474-32bf53713cf5" containerID="3579f1d8f85e742d8f3051d9ab2cc486b0cd4fdc99124c1454343e4bda4a8925" exitCode=2 Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.471606 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5mzr5" event={"ID":"6371ac18-2361-43bb-b474-32bf53713cf5","Type":"ContainerDied","Data":"3579f1d8f85e742d8f3051d9ab2cc486b0cd4fdc99124c1454343e4bda4a8925"} Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.471645 4715 scope.go:117] "RemoveContainer" containerID="2c270c4471c3151c527abc29754c047177ad269eb22e58b93dae45fcfe0bc65d" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.472062 4715 scope.go:117] "RemoveContainer" containerID="3579f1d8f85e742d8f3051d9ab2cc486b0cd4fdc99124c1454343e4bda4a8925" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.472364 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5mzr5_openshift-multus(6371ac18-2361-43bb-b474-32bf53713cf5)\"" pod="openshift-multus/multus-5mzr5" podUID="6371ac18-2361-43bb-b474-32bf53713cf5" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.736576 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/3.log" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.739108 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovn-acl-logging/0.log" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.739586 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovn-controller/0.log" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.739980 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.773906 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/02385935-e23a-4708-b57f-fd393d05651b-ovn-node-metrics-cert\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.773961 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-slash\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.773998 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-netns\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774018 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-openvswitch\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774033 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-netd\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-script-lib\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774093 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpqs9\" (UniqueName: \"kubernetes.io/projected/02385935-e23a-4708-b57f-fd393d05651b-kube-api-access-kpqs9\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774105 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-bin\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774120 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-config\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774141 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-ovn\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774156 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-node-log\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774175 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-log-socket\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774224 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-etc-openvswitch\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774242 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774260 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-systemd\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774279 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-kubelet\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774296 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-var-lib-openvswitch\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774331 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-ovn-kubernetes\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774350 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-systemd-units\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774386 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-env-overrides\") pod \"02385935-e23a-4708-b57f-fd393d05651b\" (UID: \"02385935-e23a-4708-b57f-fd393d05651b\") " Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774439 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774483 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774510 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774506 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-slash" (OuterVolumeSpecName: "host-slash") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774535 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774661 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774667 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774715 4715 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.774731 4715 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775030 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-node-log" (OuterVolumeSpecName: "node-log") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775036 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-log-socket" (OuterVolumeSpecName: "log-socket") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775063 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775080 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775085 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775083 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775113 4715 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775260 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775495 4715 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775564 4715 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775621 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775770 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.775808 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.779812 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02385935-e23a-4708-b57f-fd393d05651b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.780215 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02385935-e23a-4708-b57f-fd393d05651b-kube-api-access-kpqs9" (OuterVolumeSpecName: "kube-api-access-kpqs9") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "kube-api-access-kpqs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.792321 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "02385935-e23a-4708-b57f-fd393d05651b" (UID: "02385935-e23a-4708-b57f-fd393d05651b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.793363 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6cffw"] Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794064 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovn-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794085 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovn-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794099 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794105 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794136 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="sbdb" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794143 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="sbdb" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794153 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="northd" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794159 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="northd" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794168 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kubecfg-setup" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794173 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kubecfg-setup" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794208 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b215c8d-689d-431d-815b-b57707fb07ba" containerName="collect-profiles" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794216 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b215c8d-689d-431d-815b-b57707fb07ba" containerName="collect-profiles" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794224 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794230 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794237 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kube-rbac-proxy-node" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794243 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kube-rbac-proxy-node" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794251 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794297 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794306 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794311 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794317 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="nbdb" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794323 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="nbdb" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794330 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794336 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794344 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovn-acl-logging" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794370 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovn-acl-logging" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794481 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794489 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="northd" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794497 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovn-acl-logging" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794505 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kube-rbac-proxy-node" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794533 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794541 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="sbdb" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794548 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="nbdb" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794554 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovn-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794559 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794565 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b215c8d-689d-431d-815b-b57707fb07ba" containerName="collect-profiles" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794571 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 12:18:59 crc kubenswrapper[4715]: E1125 12:18:59.794695 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794703 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.794801 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.795028 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="02385935-e23a-4708-b57f-fd393d05651b" containerName="ovnkube-controller" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.797517 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877357 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-node-log\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877414 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-systemd-units\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877461 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-run-ovn\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877489 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-kubelet\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877512 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-run-netns\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877567 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/707ddf8a-2293-44f7-a9bb-cc3316134b62-ovn-node-metrics-cert\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877604 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-run-ovn-kubernetes\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877637 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-run-systemd\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877668 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877712 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gp5r\" (UniqueName: \"kubernetes.io/projected/707ddf8a-2293-44f7-a9bb-cc3316134b62-kube-api-access-4gp5r\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877734 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-cni-netd\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877773 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/707ddf8a-2293-44f7-a9bb-cc3316134b62-ovnkube-script-lib\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877796 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-slash\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877831 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-log-socket\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877870 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-var-lib-openvswitch\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877919 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-etc-openvswitch\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877946 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-run-openvswitch\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.877979 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/707ddf8a-2293-44f7-a9bb-cc3316134b62-env-overrides\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878019 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-cni-bin\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878039 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/707ddf8a-2293-44f7-a9bb-cc3316134b62-ovnkube-config\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878092 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878107 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/02385935-e23a-4708-b57f-fd393d05651b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878119 4715 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878132 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878144 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpqs9\" (UniqueName: \"kubernetes.io/projected/02385935-e23a-4708-b57f-fd393d05651b-kube-api-access-kpqs9\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878156 4715 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878167 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/02385935-e23a-4708-b57f-fd393d05651b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878181 4715 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878210 4715 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878222 4715 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878233 4715 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878245 4715 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878256 4715 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878282 4715 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.878294 4715 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/02385935-e23a-4708-b57f-fd393d05651b-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979136 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-node-log\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979213 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-systemd-units\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979244 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-run-ovn\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979264 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-kubelet\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979279 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-run-netns\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979287 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-node-log\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979298 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/707ddf8a-2293-44f7-a9bb-cc3316134b62-ovn-node-metrics-cert\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979427 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-run-netns\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979445 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-run-ovn-kubernetes\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979492 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-run-ovn-kubernetes\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979427 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-kubelet\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979523 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-run-systemd\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979571 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-run-systemd\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979585 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-systemd-units\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979856 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-run-ovn\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979614 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.979955 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gp5r\" (UniqueName: \"kubernetes.io/projected/707ddf8a-2293-44f7-a9bb-cc3316134b62-kube-api-access-4gp5r\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.980038 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-cni-netd\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.980105 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-cni-netd\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.980175 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-slash\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.980244 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-slash\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.980259 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/707ddf8a-2293-44f7-a9bb-cc3316134b62-ovnkube-script-lib\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.980935 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-log-socket\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981004 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-var-lib-openvswitch\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981058 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-log-socket\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981079 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-etc-openvswitch\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981105 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-var-lib-openvswitch\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981122 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-run-openvswitch\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981138 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-etc-openvswitch\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/707ddf8a-2293-44f7-a9bb-cc3316134b62-env-overrides\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981174 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-run-openvswitch\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981347 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/707ddf8a-2293-44f7-a9bb-cc3316134b62-ovnkube-script-lib\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981368 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-cni-bin\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981406 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/707ddf8a-2293-44f7-a9bb-cc3316134b62-ovnkube-config\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.981565 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/707ddf8a-2293-44f7-a9bb-cc3316134b62-host-cni-bin\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.982146 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/707ddf8a-2293-44f7-a9bb-cc3316134b62-env-overrides\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.983478 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/707ddf8a-2293-44f7-a9bb-cc3316134b62-ovnkube-config\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.984170 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/707ddf8a-2293-44f7-a9bb-cc3316134b62-ovn-node-metrics-cert\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:18:59 crc kubenswrapper[4715]: I1125 12:18:59.999306 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gp5r\" (UniqueName: \"kubernetes.io/projected/707ddf8a-2293-44f7-a9bb-cc3316134b62-kube-api-access-4gp5r\") pod \"ovnkube-node-6cffw\" (UID: \"707ddf8a-2293-44f7-a9bb-cc3316134b62\") " pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.111739 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.483120 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovnkube-controller/3.log" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.486550 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovn-acl-logging/0.log" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487225 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tzl95_02385935-e23a-4708-b57f-fd393d05651b/ovn-controller/0.log" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487756 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e" exitCode=0 Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487779 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15" exitCode=0 Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487787 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e" exitCode=0 Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487820 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e" exitCode=0 Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487832 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772" exitCode=0 Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487842 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31" exitCode=0 Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487851 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149" exitCode=143 Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487861 4715 generic.go:334] "Generic (PLEG): container finished" podID="02385935-e23a-4708-b57f-fd393d05651b" containerID="2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3" exitCode=143 Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487878 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487871 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487983 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.487997 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488010 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488031 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488047 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488013 4715 scope.go:117] "RemoveContainer" containerID="40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488131 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488161 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488170 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488178 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488210 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488218 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488227 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488235 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488241 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488255 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488270 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488280 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488287 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488295 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488302 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488308 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488316 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488324 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488332 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488339 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488349 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488361 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488372 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488381 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488389 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488398 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488407 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488415 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488423 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488430 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488436 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488446 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tzl95" event={"ID":"02385935-e23a-4708-b57f-fd393d05651b","Type":"ContainerDied","Data":"4c6fdc92b4855d79d0cf517aa6acd2eec1a1653fa1df8494fa29c67e46a3a965"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488462 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488472 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488479 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488487 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488493 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488501 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488508 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488516 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488523 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.488530 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.492062 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/2.log" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.497099 4715 generic.go:334] "Generic (PLEG): container finished" podID="707ddf8a-2293-44f7-a9bb-cc3316134b62" containerID="0e751e070a913990e5c1b0b7f2d2def740882800468c569b508ea9bace328f0f" exitCode=0 Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.497206 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerDied","Data":"0e751e070a913990e5c1b0b7f2d2def740882800468c569b508ea9bace328f0f"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.497282 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerStarted","Data":"125fea5e37b889b3ae662ef963db95f45ff96bb3560666e238aeb5308ae51800"} Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.558011 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tzl95"] Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.560705 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.566862 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tzl95"] Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.577968 4715 scope.go:117] "RemoveContainer" containerID="543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.610461 4715 scope.go:117] "RemoveContainer" containerID="d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.630149 4715 scope.go:117] "RemoveContainer" containerID="6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.655214 4715 scope.go:117] "RemoveContainer" containerID="29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.668251 4715 scope.go:117] "RemoveContainer" containerID="8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.680929 4715 scope.go:117] "RemoveContainer" containerID="3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.691465 4715 scope.go:117] "RemoveContainer" containerID="2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.699807 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02385935-e23a-4708-b57f-fd393d05651b" path="/var/lib/kubelet/pods/02385935-e23a-4708-b57f-fd393d05651b/volumes" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.706977 4715 scope.go:117] "RemoveContainer" containerID="9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.730293 4715 scope.go:117] "RemoveContainer" containerID="40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.731134 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": container with ID starting with 40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e not found: ID does not exist" containerID="40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.731211 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} err="failed to get container status \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": rpc error: code = NotFound desc = could not find container \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": container with ID starting with 40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.731249 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.731847 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\": container with ID starting with eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e not found: ID does not exist" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.731895 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} err="failed to get container status \"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\": rpc error: code = NotFound desc = could not find container \"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\": container with ID starting with eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.731932 4715 scope.go:117] "RemoveContainer" containerID="543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.732404 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\": container with ID starting with 543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15 not found: ID does not exist" containerID="543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.732441 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} err="failed to get container status \"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\": rpc error: code = NotFound desc = could not find container \"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\": container with ID starting with 543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.732461 4715 scope.go:117] "RemoveContainer" containerID="d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.732990 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\": container with ID starting with d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e not found: ID does not exist" containerID="d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.733046 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} err="failed to get container status \"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\": rpc error: code = NotFound desc = could not find container \"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\": container with ID starting with d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.733090 4715 scope.go:117] "RemoveContainer" containerID="6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.733494 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\": container with ID starting with 6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e not found: ID does not exist" containerID="6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.733526 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} err="failed to get container status \"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\": rpc error: code = NotFound desc = could not find container \"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\": container with ID starting with 6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.733548 4715 scope.go:117] "RemoveContainer" containerID="29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.733872 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\": container with ID starting with 29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772 not found: ID does not exist" containerID="29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.733915 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} err="failed to get container status \"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\": rpc error: code = NotFound desc = could not find container \"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\": container with ID starting with 29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.733944 4715 scope.go:117] "RemoveContainer" containerID="8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.734315 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\": container with ID starting with 8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31 not found: ID does not exist" containerID="8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.734354 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} err="failed to get container status \"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\": rpc error: code = NotFound desc = could not find container \"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\": container with ID starting with 8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.734378 4715 scope.go:117] "RemoveContainer" containerID="3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.734853 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\": container with ID starting with 3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149 not found: ID does not exist" containerID="3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.734888 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} err="failed to get container status \"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\": rpc error: code = NotFound desc = could not find container \"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\": container with ID starting with 3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.734909 4715 scope.go:117] "RemoveContainer" containerID="2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.735271 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\": container with ID starting with 2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3 not found: ID does not exist" containerID="2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.735301 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} err="failed to get container status \"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\": rpc error: code = NotFound desc = could not find container \"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\": container with ID starting with 2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.735347 4715 scope.go:117] "RemoveContainer" containerID="9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c" Nov 25 12:19:00 crc kubenswrapper[4715]: E1125 12:19:00.735850 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\": container with ID starting with 9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c not found: ID does not exist" containerID="9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.735904 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c"} err="failed to get container status \"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\": rpc error: code = NotFound desc = could not find container \"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\": container with ID starting with 9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.735944 4715 scope.go:117] "RemoveContainer" containerID="40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.736303 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} err="failed to get container status \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": rpc error: code = NotFound desc = could not find container \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": container with ID starting with 40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.736342 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.736689 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} err="failed to get container status \"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\": rpc error: code = NotFound desc = could not find container \"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\": container with ID starting with eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.736712 4715 scope.go:117] "RemoveContainer" containerID="543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.737006 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} err="failed to get container status \"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\": rpc error: code = NotFound desc = could not find container \"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\": container with ID starting with 543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.737040 4715 scope.go:117] "RemoveContainer" containerID="d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.737403 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} err="failed to get container status \"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\": rpc error: code = NotFound desc = could not find container \"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\": container with ID starting with d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.737432 4715 scope.go:117] "RemoveContainer" containerID="6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.737737 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} err="failed to get container status \"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\": rpc error: code = NotFound desc = could not find container \"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\": container with ID starting with 6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.737768 4715 scope.go:117] "RemoveContainer" containerID="29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.738061 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} err="failed to get container status \"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\": rpc error: code = NotFound desc = could not find container \"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\": container with ID starting with 29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.738091 4715 scope.go:117] "RemoveContainer" containerID="8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.738416 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} err="failed to get container status \"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\": rpc error: code = NotFound desc = could not find container \"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\": container with ID starting with 8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.738453 4715 scope.go:117] "RemoveContainer" containerID="3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.738743 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} err="failed to get container status \"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\": rpc error: code = NotFound desc = could not find container \"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\": container with ID starting with 3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.740166 4715 scope.go:117] "RemoveContainer" containerID="2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.740651 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} err="failed to get container status \"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\": rpc error: code = NotFound desc = could not find container \"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\": container with ID starting with 2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.740682 4715 scope.go:117] "RemoveContainer" containerID="9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.741391 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c"} err="failed to get container status \"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\": rpc error: code = NotFound desc = could not find container \"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\": container with ID starting with 9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.741457 4715 scope.go:117] "RemoveContainer" containerID="40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.741816 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} err="failed to get container status \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": rpc error: code = NotFound desc = could not find container \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": container with ID starting with 40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.741862 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.742275 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} err="failed to get container status \"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\": rpc error: code = NotFound desc = could not find container \"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\": container with ID starting with eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.742320 4715 scope.go:117] "RemoveContainer" containerID="543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.742658 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} err="failed to get container status \"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\": rpc error: code = NotFound desc = could not find container \"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\": container with ID starting with 543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.742685 4715 scope.go:117] "RemoveContainer" containerID="d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.742982 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} err="failed to get container status \"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\": rpc error: code = NotFound desc = could not find container \"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\": container with ID starting with d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.743011 4715 scope.go:117] "RemoveContainer" containerID="6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.743293 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} err="failed to get container status \"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\": rpc error: code = NotFound desc = could not find container \"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\": container with ID starting with 6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.743319 4715 scope.go:117] "RemoveContainer" containerID="29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.743606 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} err="failed to get container status \"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\": rpc error: code = NotFound desc = could not find container \"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\": container with ID starting with 29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.743639 4715 scope.go:117] "RemoveContainer" containerID="8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.743886 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} err="failed to get container status \"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\": rpc error: code = NotFound desc = could not find container \"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\": container with ID starting with 8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.743911 4715 scope.go:117] "RemoveContainer" containerID="3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.744166 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} err="failed to get container status \"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\": rpc error: code = NotFound desc = could not find container \"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\": container with ID starting with 3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.744242 4715 scope.go:117] "RemoveContainer" containerID="2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.744566 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} err="failed to get container status \"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\": rpc error: code = NotFound desc = could not find container \"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\": container with ID starting with 2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.744588 4715 scope.go:117] "RemoveContainer" containerID="9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.744822 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c"} err="failed to get container status \"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\": rpc error: code = NotFound desc = could not find container \"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\": container with ID starting with 9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.744843 4715 scope.go:117] "RemoveContainer" containerID="40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.745055 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} err="failed to get container status \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": rpc error: code = NotFound desc = could not find container \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": container with ID starting with 40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.745080 4715 scope.go:117] "RemoveContainer" containerID="eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.745414 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e"} err="failed to get container status \"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\": rpc error: code = NotFound desc = could not find container \"eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e\": container with ID starting with eeb3fa8597a062c05380687ef16d7bed0e6a46b17a55d7eac1304d1b633be40e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.745433 4715 scope.go:117] "RemoveContainer" containerID="543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.745875 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15"} err="failed to get container status \"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\": rpc error: code = NotFound desc = could not find container \"543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15\": container with ID starting with 543a47d428f7ed180f1042c0f1ad5f40967a51205ae0e46fdbca19e609676b15 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.745913 4715 scope.go:117] "RemoveContainer" containerID="d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.746237 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e"} err="failed to get container status \"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\": rpc error: code = NotFound desc = could not find container \"d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e\": container with ID starting with d6ab5d2468df5774496bad584c726ecd332be8a7a11d57e756f766118711f96e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.746266 4715 scope.go:117] "RemoveContainer" containerID="6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.746607 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e"} err="failed to get container status \"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\": rpc error: code = NotFound desc = could not find container \"6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e\": container with ID starting with 6c0825976e6816fab3d4905a4f3a968cd5e77ec94a00f0bf865a3b17173f1d1e not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.746647 4715 scope.go:117] "RemoveContainer" containerID="29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.746916 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772"} err="failed to get container status \"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\": rpc error: code = NotFound desc = could not find container \"29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772\": container with ID starting with 29f69f9f699ca39b80d163f6103b4041d18a1d73f92f08b0151b86de185c3772 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.746947 4715 scope.go:117] "RemoveContainer" containerID="8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.747223 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31"} err="failed to get container status \"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\": rpc error: code = NotFound desc = could not find container \"8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31\": container with ID starting with 8f3b353c9c1ca38d623723e3212ac7b8b45fa2121ea291ada0fd315110ab9b31 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.747249 4715 scope.go:117] "RemoveContainer" containerID="3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.747677 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149"} err="failed to get container status \"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\": rpc error: code = NotFound desc = could not find container \"3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149\": container with ID starting with 3e41544570147445a3160cc9b549a98e4244c1af5d14eb610ce9f0620f837149 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.747705 4715 scope.go:117] "RemoveContainer" containerID="2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.747999 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3"} err="failed to get container status \"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\": rpc error: code = NotFound desc = could not find container \"2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3\": container with ID starting with 2f411d85f30bb649b926a9f9c2a7eed7e060ae30f1f46976a6eb57d49ad064d3 not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.748025 4715 scope.go:117] "RemoveContainer" containerID="9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.748377 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c"} err="failed to get container status \"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\": rpc error: code = NotFound desc = could not find container \"9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c\": container with ID starting with 9510547db15ebb983c1aa5940842efffcb806a7eabcc5d72a99a5ab524055f9c not found: ID does not exist" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.748400 4715 scope.go:117] "RemoveContainer" containerID="40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e" Nov 25 12:19:00 crc kubenswrapper[4715]: I1125 12:19:00.748679 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e"} err="failed to get container status \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": rpc error: code = NotFound desc = could not find container \"40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e\": container with ID starting with 40b2c2c91e466472f3840ff48b618fe076d43f5fa039e4bbc8273684d507c87e not found: ID does not exist" Nov 25 12:19:01 crc kubenswrapper[4715]: I1125 12:19:01.511740 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerStarted","Data":"316a75862c495dbe54342adc61f1020450791fd6500e18cbd0759eead4097f04"} Nov 25 12:19:01 crc kubenswrapper[4715]: I1125 12:19:01.512395 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerStarted","Data":"96585782c507c07c01f4b3578a0287c07980aa08cab1965fe47bebf29643a653"} Nov 25 12:19:01 crc kubenswrapper[4715]: I1125 12:19:01.512478 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerStarted","Data":"e0043eae3cab114cef37822b86d657f8f24e7159c13bcdcc8471fe6185a144bc"} Nov 25 12:19:01 crc kubenswrapper[4715]: I1125 12:19:01.512509 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerStarted","Data":"0d4e2060b9aa4cd4ce3fb2669ecd6b96dcd0d44e24f1b35ac6f95861f5e8195e"} Nov 25 12:19:02 crc kubenswrapper[4715]: I1125 12:19:02.523398 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerStarted","Data":"53ed6f6368a0e1aa0b493f57d77daf283d4bbd37dd5140735d935b4d8b187ad0"} Nov 25 12:19:02 crc kubenswrapper[4715]: I1125 12:19:02.523797 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerStarted","Data":"beda22b220eac9bd9c2ba1e0708d0b0372df35d509af465337af79dcd651e245"} Nov 25 12:19:04 crc kubenswrapper[4715]: I1125 12:19:04.539491 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerStarted","Data":"41a34cff099283f9b747bc0f6f19fe2c8d3003216e1595a5453e098b044f1a22"} Nov 25 12:19:06 crc kubenswrapper[4715]: I1125 12:19:06.553906 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" event={"ID":"707ddf8a-2293-44f7-a9bb-cc3316134b62","Type":"ContainerStarted","Data":"27b43b9aa18e57ab63bcac92c32f2e2850189102f0154dcc7795f945ed2d382e"} Nov 25 12:19:06 crc kubenswrapper[4715]: I1125 12:19:06.554146 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:19:06 crc kubenswrapper[4715]: I1125 12:19:06.554159 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:19:06 crc kubenswrapper[4715]: I1125 12:19:06.581205 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:19:06 crc kubenswrapper[4715]: I1125 12:19:06.588470 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" podStartSLOduration=7.588454561 podStartE2EDuration="7.588454561s" podCreationTimestamp="2025-11-25 12:18:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:19:06.585223448 +0000 UTC m=+597.092726469" watchObservedRunningTime="2025-11-25 12:19:06.588454561 +0000 UTC m=+597.095957582" Nov 25 12:19:07 crc kubenswrapper[4715]: I1125 12:19:07.558562 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:19:07 crc kubenswrapper[4715]: I1125 12:19:07.585816 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:19:14 crc kubenswrapper[4715]: I1125 12:19:14.694097 4715 scope.go:117] "RemoveContainer" containerID="3579f1d8f85e742d8f3051d9ab2cc486b0cd4fdc99124c1454343e4bda4a8925" Nov 25 12:19:14 crc kubenswrapper[4715]: E1125 12:19:14.694878 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5mzr5_openshift-multus(6371ac18-2361-43bb-b474-32bf53713cf5)\"" pod="openshift-multus/multus-5mzr5" podUID="6371ac18-2361-43bb-b474-32bf53713cf5" Nov 25 12:19:17 crc kubenswrapper[4715]: I1125 12:19:17.944265 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:19:17 crc kubenswrapper[4715]: I1125 12:19:17.944605 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.813948 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5"] Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.815679 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.817875 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.824576 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5"] Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.878791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.878873 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.878898 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mflhr\" (UniqueName: \"kubernetes.io/projected/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-kube-api-access-mflhr\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.981009 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.981096 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.981133 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mflhr\" (UniqueName: \"kubernetes.io/projected/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-kube-api-access-mflhr\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.981652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:22 crc kubenswrapper[4715]: I1125 12:19:22.981652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:23 crc kubenswrapper[4715]: I1125 12:19:23.003631 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mflhr\" (UniqueName: \"kubernetes.io/projected/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-kube-api-access-mflhr\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:23 crc kubenswrapper[4715]: I1125 12:19:23.132019 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:23 crc kubenswrapper[4715]: E1125 12:19:23.162305 4715 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455_0(f49f5a9ec376809adc41cca12c192f9d1e38d6d7ea010be8ed7fd4393c70ce09): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 12:19:23 crc kubenswrapper[4715]: E1125 12:19:23.162391 4715 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455_0(f49f5a9ec376809adc41cca12c192f9d1e38d6d7ea010be8ed7fd4393c70ce09): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:23 crc kubenswrapper[4715]: E1125 12:19:23.162418 4715 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455_0(f49f5a9ec376809adc41cca12c192f9d1e38d6d7ea010be8ed7fd4393c70ce09): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:23 crc kubenswrapper[4715]: E1125 12:19:23.162480 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace(ac0313a0-ef82-409e-8ceb-fa5ccb8b9455)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace(ac0313a0-ef82-409e-8ceb-fa5ccb8b9455)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455_0(f49f5a9ec376809adc41cca12c192f9d1e38d6d7ea010be8ed7fd4393c70ce09): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" podUID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" Nov 25 12:19:23 crc kubenswrapper[4715]: I1125 12:19:23.656673 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:23 crc kubenswrapper[4715]: I1125 12:19:23.657297 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:23 crc kubenswrapper[4715]: E1125 12:19:23.683212 4715 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455_0(c47e602707294c9d5e083a9c809dca1aef9c8a5515a0d87605229b40444f92e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 12:19:23 crc kubenswrapper[4715]: E1125 12:19:23.683313 4715 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455_0(c47e602707294c9d5e083a9c809dca1aef9c8a5515a0d87605229b40444f92e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:23 crc kubenswrapper[4715]: E1125 12:19:23.683351 4715 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455_0(c47e602707294c9d5e083a9c809dca1aef9c8a5515a0d87605229b40444f92e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:23 crc kubenswrapper[4715]: E1125 12:19:23.683426 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace(ac0313a0-ef82-409e-8ceb-fa5ccb8b9455)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace(ac0313a0-ef82-409e-8ceb-fa5ccb8b9455)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_openshift-marketplace_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455_0(c47e602707294c9d5e083a9c809dca1aef9c8a5515a0d87605229b40444f92e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" podUID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" Nov 25 12:19:25 crc kubenswrapper[4715]: I1125 12:19:25.694864 4715 scope.go:117] "RemoveContainer" containerID="3579f1d8f85e742d8f3051d9ab2cc486b0cd4fdc99124c1454343e4bda4a8925" Nov 25 12:19:26 crc kubenswrapper[4715]: I1125 12:19:26.675112 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5mzr5_6371ac18-2361-43bb-b474-32bf53713cf5/kube-multus/2.log" Nov 25 12:19:26 crc kubenswrapper[4715]: I1125 12:19:26.675165 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5mzr5" event={"ID":"6371ac18-2361-43bb-b474-32bf53713cf5","Type":"ContainerStarted","Data":"b438c15ac4674e59cfc77176f250fc0415b1735b53751b28abf41ca0f66fba1b"} Nov 25 12:19:30 crc kubenswrapper[4715]: I1125 12:19:30.155317 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6cffw" Nov 25 12:19:36 crc kubenswrapper[4715]: I1125 12:19:36.693076 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:36 crc kubenswrapper[4715]: I1125 12:19:36.694088 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:36 crc kubenswrapper[4715]: I1125 12:19:36.920127 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5"] Nov 25 12:19:36 crc kubenswrapper[4715]: W1125 12:19:36.926212 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac0313a0_ef82_409e_8ceb_fa5ccb8b9455.slice/crio-0be06137e9fac09ed81bcf1a1f5c8d7e31e69d49fafc635632882f5a0d2451fa WatchSource:0}: Error finding container 0be06137e9fac09ed81bcf1a1f5c8d7e31e69d49fafc635632882f5a0d2451fa: Status 404 returned error can't find the container with id 0be06137e9fac09ed81bcf1a1f5c8d7e31e69d49fafc635632882f5a0d2451fa Nov 25 12:19:37 crc kubenswrapper[4715]: I1125 12:19:37.742332 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" event={"ID":"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455","Type":"ContainerStarted","Data":"57992614bc376d2ecef5292a7fa994bb6e5410b5af0add0e9a1b2cce7f04b5ec"} Nov 25 12:19:37 crc kubenswrapper[4715]: I1125 12:19:37.742816 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" event={"ID":"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455","Type":"ContainerStarted","Data":"0be06137e9fac09ed81bcf1a1f5c8d7e31e69d49fafc635632882f5a0d2451fa"} Nov 25 12:19:39 crc kubenswrapper[4715]: I1125 12:19:39.759992 4715 generic.go:334] "Generic (PLEG): container finished" podID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerID="57992614bc376d2ecef5292a7fa994bb6e5410b5af0add0e9a1b2cce7f04b5ec" exitCode=0 Nov 25 12:19:39 crc kubenswrapper[4715]: I1125 12:19:39.760255 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" event={"ID":"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455","Type":"ContainerDied","Data":"57992614bc376d2ecef5292a7fa994bb6e5410b5af0add0e9a1b2cce7f04b5ec"} Nov 25 12:19:39 crc kubenswrapper[4715]: I1125 12:19:39.763502 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 12:19:41 crc kubenswrapper[4715]: I1125 12:19:41.772650 4715 generic.go:334] "Generic (PLEG): container finished" podID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerID="dd9126bd8860a0cada0462b844948041ea606bb0137d81b21f8ce40290c798cf" exitCode=0 Nov 25 12:19:41 crc kubenswrapper[4715]: I1125 12:19:41.772737 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" event={"ID":"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455","Type":"ContainerDied","Data":"dd9126bd8860a0cada0462b844948041ea606bb0137d81b21f8ce40290c798cf"} Nov 25 12:19:42 crc kubenswrapper[4715]: I1125 12:19:42.782446 4715 generic.go:334] "Generic (PLEG): container finished" podID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerID="c27a753cd8d7018e543bc75481a3001f811fd7d3d3b2a1bfbc874f4fb6bcac32" exitCode=0 Nov 25 12:19:42 crc kubenswrapper[4715]: I1125 12:19:42.782495 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" event={"ID":"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455","Type":"ContainerDied","Data":"c27a753cd8d7018e543bc75481a3001f811fd7d3d3b2a1bfbc874f4fb6bcac32"} Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.026498 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.172635 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-util\") pod \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.172734 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-bundle\") pod \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.172802 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mflhr\" (UniqueName: \"kubernetes.io/projected/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-kube-api-access-mflhr\") pod \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\" (UID: \"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455\") " Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.173949 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-bundle" (OuterVolumeSpecName: "bundle") pod "ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" (UID: "ac0313a0-ef82-409e-8ceb-fa5ccb8b9455"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.178343 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-kube-api-access-mflhr" (OuterVolumeSpecName: "kube-api-access-mflhr") pod "ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" (UID: "ac0313a0-ef82-409e-8ceb-fa5ccb8b9455"). InnerVolumeSpecName "kube-api-access-mflhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.189937 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-util" (OuterVolumeSpecName: "util") pod "ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" (UID: "ac0313a0-ef82-409e-8ceb-fa5ccb8b9455"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.274312 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-util\") on node \"crc\" DevicePath \"\"" Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.274360 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.274372 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mflhr\" (UniqueName: \"kubernetes.io/projected/ac0313a0-ef82-409e-8ceb-fa5ccb8b9455-kube-api-access-mflhr\") on node \"crc\" DevicePath \"\"" Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.793448 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" event={"ID":"ac0313a0-ef82-409e-8ceb-fa5ccb8b9455","Type":"ContainerDied","Data":"0be06137e9fac09ed81bcf1a1f5c8d7e31e69d49fafc635632882f5a0d2451fa"} Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.793753 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0be06137e9fac09ed81bcf1a1f5c8d7e31e69d49fafc635632882f5a0d2451fa" Nov 25 12:19:44 crc kubenswrapper[4715]: I1125 12:19:44.793518 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5" Nov 25 12:19:47 crc kubenswrapper[4715]: I1125 12:19:47.943294 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:19:47 crc kubenswrapper[4715]: I1125 12:19:47.943366 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:19:47 crc kubenswrapper[4715]: I1125 12:19:47.943412 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:19:47 crc kubenswrapper[4715]: I1125 12:19:47.943896 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d16257d5ac23a2866262558b39c2b844ea7649b0c06fa81698cc8fe1fd3c8f95"} pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:19:47 crc kubenswrapper[4715]: I1125 12:19:47.943949 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" containerID="cri-o://d16257d5ac23a2866262558b39c2b844ea7649b0c06fa81698cc8fe1fd3c8f95" gracePeriod=600 Nov 25 12:19:48 crc kubenswrapper[4715]: I1125 12:19:48.821701 4715 generic.go:334] "Generic (PLEG): container finished" podID="51676837-adc4-4424-b527-920a6528b6a2" containerID="d16257d5ac23a2866262558b39c2b844ea7649b0c06fa81698cc8fe1fd3c8f95" exitCode=0 Nov 25 12:19:48 crc kubenswrapper[4715]: I1125 12:19:48.821768 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerDied","Data":"d16257d5ac23a2866262558b39c2b844ea7649b0c06fa81698cc8fe1fd3c8f95"} Nov 25 12:19:48 crc kubenswrapper[4715]: I1125 12:19:48.822303 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"23a04d777596661d2024d01da087796f35f637e5c71480856ee076f6f5b38b7a"} Nov 25 12:19:48 crc kubenswrapper[4715]: I1125 12:19:48.822351 4715 scope.go:117] "RemoveContainer" containerID="133f6a3b52123e9fe263e667ae6646c6fa53439cce70b559acdf67b4d86c9ee6" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.092006 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw"] Nov 25 12:19:56 crc kubenswrapper[4715]: E1125 12:19:56.092788 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerName="util" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.092806 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerName="util" Nov 25 12:19:56 crc kubenswrapper[4715]: E1125 12:19:56.092822 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerName="extract" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.092831 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerName="extract" Nov 25 12:19:56 crc kubenswrapper[4715]: E1125 12:19:56.092847 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerName="pull" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.092856 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerName="pull" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.092961 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac0313a0-ef82-409e-8ceb-fa5ccb8b9455" containerName="extract" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.093440 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.096106 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.096294 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.096465 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.096946 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zl6db" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.101303 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.118066 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw"] Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.217564 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/691af904-3d82-40e2-946f-2c7d3fd237a2-webhook-cert\") pod \"metallb-operator-controller-manager-6cb7984d4b-5mkxw\" (UID: \"691af904-3d82-40e2-946f-2c7d3fd237a2\") " pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.217623 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4v9m\" (UniqueName: \"kubernetes.io/projected/691af904-3d82-40e2-946f-2c7d3fd237a2-kube-api-access-g4v9m\") pod \"metallb-operator-controller-manager-6cb7984d4b-5mkxw\" (UID: \"691af904-3d82-40e2-946f-2c7d3fd237a2\") " pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.217670 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/691af904-3d82-40e2-946f-2c7d3fd237a2-apiservice-cert\") pod \"metallb-operator-controller-manager-6cb7984d4b-5mkxw\" (UID: \"691af904-3d82-40e2-946f-2c7d3fd237a2\") " pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.319096 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/691af904-3d82-40e2-946f-2c7d3fd237a2-webhook-cert\") pod \"metallb-operator-controller-manager-6cb7984d4b-5mkxw\" (UID: \"691af904-3d82-40e2-946f-2c7d3fd237a2\") " pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.319147 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4v9m\" (UniqueName: \"kubernetes.io/projected/691af904-3d82-40e2-946f-2c7d3fd237a2-kube-api-access-g4v9m\") pod \"metallb-operator-controller-manager-6cb7984d4b-5mkxw\" (UID: \"691af904-3d82-40e2-946f-2c7d3fd237a2\") " pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.319169 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/691af904-3d82-40e2-946f-2c7d3fd237a2-apiservice-cert\") pod \"metallb-operator-controller-manager-6cb7984d4b-5mkxw\" (UID: \"691af904-3d82-40e2-946f-2c7d3fd237a2\") " pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.325931 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/691af904-3d82-40e2-946f-2c7d3fd237a2-apiservice-cert\") pod \"metallb-operator-controller-manager-6cb7984d4b-5mkxw\" (UID: \"691af904-3d82-40e2-946f-2c7d3fd237a2\") " pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.325933 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/691af904-3d82-40e2-946f-2c7d3fd237a2-webhook-cert\") pod \"metallb-operator-controller-manager-6cb7984d4b-5mkxw\" (UID: \"691af904-3d82-40e2-946f-2c7d3fd237a2\") " pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.356946 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4v9m\" (UniqueName: \"kubernetes.io/projected/691af904-3d82-40e2-946f-2c7d3fd237a2-kube-api-access-g4v9m\") pod \"metallb-operator-controller-manager-6cb7984d4b-5mkxw\" (UID: \"691af904-3d82-40e2-946f-2c7d3fd237a2\") " pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.409831 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.454678 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-84945578d6-hsb65"] Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.455499 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.458540 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.459173 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.460614 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-wxqdv" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.471727 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84945578d6-hsb65"] Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.521690 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/449e1cf8-4b25-4371-8e3c-505f09548785-webhook-cert\") pod \"metallb-operator-webhook-server-84945578d6-hsb65\" (UID: \"449e1cf8-4b25-4371-8e3c-505f09548785\") " pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.521749 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz29p\" (UniqueName: \"kubernetes.io/projected/449e1cf8-4b25-4371-8e3c-505f09548785-kube-api-access-hz29p\") pod \"metallb-operator-webhook-server-84945578d6-hsb65\" (UID: \"449e1cf8-4b25-4371-8e3c-505f09548785\") " pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.521804 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/449e1cf8-4b25-4371-8e3c-505f09548785-apiservice-cert\") pod \"metallb-operator-webhook-server-84945578d6-hsb65\" (UID: \"449e1cf8-4b25-4371-8e3c-505f09548785\") " pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.622621 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz29p\" (UniqueName: \"kubernetes.io/projected/449e1cf8-4b25-4371-8e3c-505f09548785-kube-api-access-hz29p\") pod \"metallb-operator-webhook-server-84945578d6-hsb65\" (UID: \"449e1cf8-4b25-4371-8e3c-505f09548785\") " pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.622707 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/449e1cf8-4b25-4371-8e3c-505f09548785-apiservice-cert\") pod \"metallb-operator-webhook-server-84945578d6-hsb65\" (UID: \"449e1cf8-4b25-4371-8e3c-505f09548785\") " pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.622729 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/449e1cf8-4b25-4371-8e3c-505f09548785-webhook-cert\") pod \"metallb-operator-webhook-server-84945578d6-hsb65\" (UID: \"449e1cf8-4b25-4371-8e3c-505f09548785\") " pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.628335 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/449e1cf8-4b25-4371-8e3c-505f09548785-apiservice-cert\") pod \"metallb-operator-webhook-server-84945578d6-hsb65\" (UID: \"449e1cf8-4b25-4371-8e3c-505f09548785\") " pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.628370 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/449e1cf8-4b25-4371-8e3c-505f09548785-webhook-cert\") pod \"metallb-operator-webhook-server-84945578d6-hsb65\" (UID: \"449e1cf8-4b25-4371-8e3c-505f09548785\") " pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.632284 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw"] Nov 25 12:19:56 crc kubenswrapper[4715]: W1125 12:19:56.633125 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod691af904_3d82_40e2_946f_2c7d3fd237a2.slice/crio-add1aa2fb4ea161bf70c66b9642985fd5fe886eab382c1e82b1634dad937ccd3 WatchSource:0}: Error finding container add1aa2fb4ea161bf70c66b9642985fd5fe886eab382c1e82b1634dad937ccd3: Status 404 returned error can't find the container with id add1aa2fb4ea161bf70c66b9642985fd5fe886eab382c1e82b1634dad937ccd3 Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.640492 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz29p\" (UniqueName: \"kubernetes.io/projected/449e1cf8-4b25-4371-8e3c-505f09548785-kube-api-access-hz29p\") pod \"metallb-operator-webhook-server-84945578d6-hsb65\" (UID: \"449e1cf8-4b25-4371-8e3c-505f09548785\") " pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.778486 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:19:56 crc kubenswrapper[4715]: I1125 12:19:56.887269 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" event={"ID":"691af904-3d82-40e2-946f-2c7d3fd237a2","Type":"ContainerStarted","Data":"add1aa2fb4ea161bf70c66b9642985fd5fe886eab382c1e82b1634dad937ccd3"} Nov 25 12:19:57 crc kubenswrapper[4715]: I1125 12:19:57.099354 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84945578d6-hsb65"] Nov 25 12:19:57 crc kubenswrapper[4715]: W1125 12:19:57.104496 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod449e1cf8_4b25_4371_8e3c_505f09548785.slice/crio-a75ce70c3aec017b9b0bb14a4dfe1e737e4c54249ead92324a422ea9caff1b8d WatchSource:0}: Error finding container a75ce70c3aec017b9b0bb14a4dfe1e737e4c54249ead92324a422ea9caff1b8d: Status 404 returned error can't find the container with id a75ce70c3aec017b9b0bb14a4dfe1e737e4c54249ead92324a422ea9caff1b8d Nov 25 12:19:57 crc kubenswrapper[4715]: I1125 12:19:57.902350 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" event={"ID":"449e1cf8-4b25-4371-8e3c-505f09548785","Type":"ContainerStarted","Data":"a75ce70c3aec017b9b0bb14a4dfe1e737e4c54249ead92324a422ea9caff1b8d"} Nov 25 12:20:01 crc kubenswrapper[4715]: I1125 12:20:01.923965 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" event={"ID":"449e1cf8-4b25-4371-8e3c-505f09548785","Type":"ContainerStarted","Data":"1f4a46a60c5ba4d9d703beecd4c19f8379814f1a921dc54aac492adf278f6874"} Nov 25 12:20:01 crc kubenswrapper[4715]: I1125 12:20:01.924556 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:20:01 crc kubenswrapper[4715]: I1125 12:20:01.925392 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" event={"ID":"691af904-3d82-40e2-946f-2c7d3fd237a2","Type":"ContainerStarted","Data":"f24acef2d436c798d308887800d6153d803c90eb1508aa4e455ca592f60f8e21"} Nov 25 12:20:01 crc kubenswrapper[4715]: I1125 12:20:01.925568 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:20:01 crc kubenswrapper[4715]: I1125 12:20:01.946242 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" podStartSLOduration=1.77695907 podStartE2EDuration="5.94622551s" podCreationTimestamp="2025-11-25 12:19:56 +0000 UTC" firstStartedPulling="2025-11-25 12:19:57.108101066 +0000 UTC m=+647.615604087" lastFinishedPulling="2025-11-25 12:20:01.277367506 +0000 UTC m=+651.784870527" observedRunningTime="2025-11-25 12:20:01.942104341 +0000 UTC m=+652.449607372" watchObservedRunningTime="2025-11-25 12:20:01.94622551 +0000 UTC m=+652.453728541" Nov 25 12:20:01 crc kubenswrapper[4715]: I1125 12:20:01.968420 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" podStartSLOduration=1.3448171389999999 podStartE2EDuration="5.96839853s" podCreationTimestamp="2025-11-25 12:19:56 +0000 UTC" firstStartedPulling="2025-11-25 12:19:56.635463636 +0000 UTC m=+647.142966657" lastFinishedPulling="2025-11-25 12:20:01.259045027 +0000 UTC m=+651.766548048" observedRunningTime="2025-11-25 12:20:01.962804968 +0000 UTC m=+652.470307999" watchObservedRunningTime="2025-11-25 12:20:01.96839853 +0000 UTC m=+652.475901571" Nov 25 12:20:16 crc kubenswrapper[4715]: I1125 12:20:16.782756 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-84945578d6-hsb65" Nov 25 12:20:36 crc kubenswrapper[4715]: I1125 12:20:36.413370 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6cb7984d4b-5mkxw" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.132483 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ddwdz"] Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.134663 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.136493 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.138801 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-p6qf7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.143287 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-8msp7"] Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.144037 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.144534 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.146527 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.151367 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-frr-conf\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.151417 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/874e4bd7-5cb5-47f2-b905-fe2dc002546b-cert\") pod \"frr-k8s-webhook-server-6998585d5-8msp7\" (UID: \"874e4bd7-5cb5-47f2-b905-fe2dc002546b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.151532 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-frr-startup\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.151584 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncjgd\" (UniqueName: \"kubernetes.io/projected/874e4bd7-5cb5-47f2-b905-fe2dc002546b-kube-api-access-ncjgd\") pod \"frr-k8s-webhook-server-6998585d5-8msp7\" (UID: \"874e4bd7-5cb5-47f2-b905-fe2dc002546b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.151623 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-frr-sockets\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.151780 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-metrics\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.151820 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-metrics-certs\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.151890 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-reloader\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.151915 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjcx5\" (UniqueName: \"kubernetes.io/projected/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-kube-api-access-wjcx5\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.165391 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-8msp7"] Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.218475 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-lf54k"] Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.219465 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.222084 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.222308 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.222929 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-9zhmq" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.233751 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.240918 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-rwp55"] Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.241732 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.243704 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.249234 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-rwp55"] Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.252922 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-frr-conf\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.252961 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/874e4bd7-5cb5-47f2-b905-fe2dc002546b-cert\") pod \"frr-k8s-webhook-server-6998585d5-8msp7\" (UID: \"874e4bd7-5cb5-47f2-b905-fe2dc002546b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.252988 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-metrics-certs\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253032 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-frr-startup\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253056 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncjgd\" (UniqueName: \"kubernetes.io/projected/874e4bd7-5cb5-47f2-b905-fe2dc002546b-kube-api-access-ncjgd\") pod \"frr-k8s-webhook-server-6998585d5-8msp7\" (UID: \"874e4bd7-5cb5-47f2-b905-fe2dc002546b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253081 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-frr-sockets\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253113 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d193542e-7b7f-4ee2-99dc-6db50e82c1c2-cert\") pod \"controller-6c7b4b5f48-rwp55\" (UID: \"d193542e-7b7f-4ee2-99dc-6db50e82c1c2\") " pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253138 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-memberlist\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253165 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgvlh\" (UniqueName: \"kubernetes.io/projected/f6028855-f32e-407c-a4d5-66e8caa6db64-kube-api-access-rgvlh\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253224 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-metrics\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253245 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-metrics-certs\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253271 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d193542e-7b7f-4ee2-99dc-6db50e82c1c2-metrics-certs\") pod \"controller-6c7b4b5f48-rwp55\" (UID: \"d193542e-7b7f-4ee2-99dc-6db50e82c1c2\") " pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253299 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f6028855-f32e-407c-a4d5-66e8caa6db64-metallb-excludel2\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253324 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p7r4\" (UniqueName: \"kubernetes.io/projected/d193542e-7b7f-4ee2-99dc-6db50e82c1c2-kube-api-access-7p7r4\") pod \"controller-6c7b4b5f48-rwp55\" (UID: \"d193542e-7b7f-4ee2-99dc-6db50e82c1c2\") " pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253350 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-reloader\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.253377 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjcx5\" (UniqueName: \"kubernetes.io/projected/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-kube-api-access-wjcx5\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.254049 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-frr-conf\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.254132 4715 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.254177 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/874e4bd7-5cb5-47f2-b905-fe2dc002546b-cert podName:874e4bd7-5cb5-47f2-b905-fe2dc002546b nodeName:}" failed. No retries permitted until 2025-11-25 12:20:37.754159456 +0000 UTC m=+688.261662487 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/874e4bd7-5cb5-47f2-b905-fe2dc002546b-cert") pod "frr-k8s-webhook-server-6998585d5-8msp7" (UID: "874e4bd7-5cb5-47f2-b905-fe2dc002546b") : secret "frr-k8s-webhook-server-cert" not found Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.254559 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-metrics\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.254629 4715 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.254658 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-metrics-certs podName:0191af85-1aa7-4c92-a45d-fc502d5dbf8c nodeName:}" failed. No retries permitted until 2025-11-25 12:20:37.75464812 +0000 UTC m=+688.262151141 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-metrics-certs") pod "frr-k8s-ddwdz" (UID: "0191af85-1aa7-4c92-a45d-fc502d5dbf8c") : secret "frr-k8s-certs-secret" not found Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.254870 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-reloader\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.254996 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-frr-startup\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.255084 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-frr-sockets\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.271522 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjcx5\" (UniqueName: \"kubernetes.io/projected/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-kube-api-access-wjcx5\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.276265 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncjgd\" (UniqueName: \"kubernetes.io/projected/874e4bd7-5cb5-47f2-b905-fe2dc002546b-kube-api-access-ncjgd\") pod \"frr-k8s-webhook-server-6998585d5-8msp7\" (UID: \"874e4bd7-5cb5-47f2-b905-fe2dc002546b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.354393 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d193542e-7b7f-4ee2-99dc-6db50e82c1c2-cert\") pod \"controller-6c7b4b5f48-rwp55\" (UID: \"d193542e-7b7f-4ee2-99dc-6db50e82c1c2\") " pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.354447 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-memberlist\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.354475 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgvlh\" (UniqueName: \"kubernetes.io/projected/f6028855-f32e-407c-a4d5-66e8caa6db64-kube-api-access-rgvlh\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.354512 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d193542e-7b7f-4ee2-99dc-6db50e82c1c2-metrics-certs\") pod \"controller-6c7b4b5f48-rwp55\" (UID: \"d193542e-7b7f-4ee2-99dc-6db50e82c1c2\") " pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.354532 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f6028855-f32e-407c-a4d5-66e8caa6db64-metallb-excludel2\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.354552 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p7r4\" (UniqueName: \"kubernetes.io/projected/d193542e-7b7f-4ee2-99dc-6db50e82c1c2-kube-api-access-7p7r4\") pod \"controller-6c7b4b5f48-rwp55\" (UID: \"d193542e-7b7f-4ee2-99dc-6db50e82c1c2\") " pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.354585 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-metrics-certs\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.354718 4715 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.354766 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-metrics-certs podName:f6028855-f32e-407c-a4d5-66e8caa6db64 nodeName:}" failed. No retries permitted until 2025-11-25 12:20:37.854751699 +0000 UTC m=+688.362254720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-metrics-certs") pod "speaker-lf54k" (UID: "f6028855-f32e-407c-a4d5-66e8caa6db64") : secret "speaker-certs-secret" not found Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.354851 4715 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.354921 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-memberlist podName:f6028855-f32e-407c-a4d5-66e8caa6db64 nodeName:}" failed. No retries permitted until 2025-11-25 12:20:37.854900843 +0000 UTC m=+688.362404074 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-memberlist") pod "speaker-lf54k" (UID: "f6028855-f32e-407c-a4d5-66e8caa6db64") : secret "metallb-memberlist" not found Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.355645 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f6028855-f32e-407c-a4d5-66e8caa6db64-metallb-excludel2\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.355778 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.359826 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d193542e-7b7f-4ee2-99dc-6db50e82c1c2-metrics-certs\") pod \"controller-6c7b4b5f48-rwp55\" (UID: \"d193542e-7b7f-4ee2-99dc-6db50e82c1c2\") " pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.369604 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d193542e-7b7f-4ee2-99dc-6db50e82c1c2-cert\") pod \"controller-6c7b4b5f48-rwp55\" (UID: \"d193542e-7b7f-4ee2-99dc-6db50e82c1c2\") " pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.372900 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p7r4\" (UniqueName: \"kubernetes.io/projected/d193542e-7b7f-4ee2-99dc-6db50e82c1c2-kube-api-access-7p7r4\") pod \"controller-6c7b4b5f48-rwp55\" (UID: \"d193542e-7b7f-4ee2-99dc-6db50e82c1c2\") " pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.375850 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgvlh\" (UniqueName: \"kubernetes.io/projected/f6028855-f32e-407c-a4d5-66e8caa6db64-kube-api-access-rgvlh\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.554726 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.748347 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-rwp55"] Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.758973 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/874e4bd7-5cb5-47f2-b905-fe2dc002546b-cert\") pod \"frr-k8s-webhook-server-6998585d5-8msp7\" (UID: \"874e4bd7-5cb5-47f2-b905-fe2dc002546b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.759100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-metrics-certs\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.762533 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/874e4bd7-5cb5-47f2-b905-fe2dc002546b-cert\") pod \"frr-k8s-webhook-server-6998585d5-8msp7\" (UID: \"874e4bd7-5cb5-47f2-b905-fe2dc002546b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.762613 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0191af85-1aa7-4c92-a45d-fc502d5dbf8c-metrics-certs\") pod \"frr-k8s-ddwdz\" (UID: \"0191af85-1aa7-4c92-a45d-fc502d5dbf8c\") " pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.774377 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.864860 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-memberlist\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.865470 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-metrics-certs\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.865124 4715 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 12:20:37 crc kubenswrapper[4715]: E1125 12:20:37.865712 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-memberlist podName:f6028855-f32e-407c-a4d5-66e8caa6db64 nodeName:}" failed. No retries permitted until 2025-11-25 12:20:38.865681945 +0000 UTC m=+689.373184986 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-memberlist") pod "speaker-lf54k" (UID: "f6028855-f32e-407c-a4d5-66e8caa6db64") : secret "metallb-memberlist" not found Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.871513 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-metrics-certs\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:37 crc kubenswrapper[4715]: I1125 12:20:37.959153 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-8msp7"] Nov 25 12:20:38 crc kubenswrapper[4715]: I1125 12:20:38.062866 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:38 crc kubenswrapper[4715]: I1125 12:20:38.129453 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" event={"ID":"874e4bd7-5cb5-47f2-b905-fe2dc002546b","Type":"ContainerStarted","Data":"c4b36bd941071e019e6c7eaa3d89bfaf15e24c37bdfca468a292faba85538830"} Nov 25 12:20:38 crc kubenswrapper[4715]: I1125 12:20:38.132520 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-rwp55" event={"ID":"d193542e-7b7f-4ee2-99dc-6db50e82c1c2","Type":"ContainerStarted","Data":"8f762b3872c8a46fba52fec61c1401abdd0432d2b5de0e3711d5697b0368e000"} Nov 25 12:20:38 crc kubenswrapper[4715]: I1125 12:20:38.132547 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-rwp55" event={"ID":"d193542e-7b7f-4ee2-99dc-6db50e82c1c2","Type":"ContainerStarted","Data":"04e9ff6ff2bf499c29cd8f53cb918b833d4b3648d37f482f55d5d17e17ad6b81"} Nov 25 12:20:38 crc kubenswrapper[4715]: I1125 12:20:38.877068 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-memberlist\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:38 crc kubenswrapper[4715]: I1125 12:20:38.882777 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f6028855-f32e-407c-a4d5-66e8caa6db64-memberlist\") pod \"speaker-lf54k\" (UID: \"f6028855-f32e-407c-a4d5-66e8caa6db64\") " pod="metallb-system/speaker-lf54k" Nov 25 12:20:39 crc kubenswrapper[4715]: I1125 12:20:39.034436 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-lf54k" Nov 25 12:20:39 crc kubenswrapper[4715]: W1125 12:20:39.063237 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6028855_f32e_407c_a4d5_66e8caa6db64.slice/crio-704de592d41447cf9c55dae47fd1edf099804933b5e6ff4865fa185377a76fb3 WatchSource:0}: Error finding container 704de592d41447cf9c55dae47fd1edf099804933b5e6ff4865fa185377a76fb3: Status 404 returned error can't find the container with id 704de592d41447cf9c55dae47fd1edf099804933b5e6ff4865fa185377a76fb3 Nov 25 12:20:39 crc kubenswrapper[4715]: I1125 12:20:39.140386 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-lf54k" event={"ID":"f6028855-f32e-407c-a4d5-66e8caa6db64","Type":"ContainerStarted","Data":"704de592d41447cf9c55dae47fd1edf099804933b5e6ff4865fa185377a76fb3"} Nov 25 12:20:39 crc kubenswrapper[4715]: I1125 12:20:39.141806 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerStarted","Data":"e1d99d12df7de80b9e244b8cb7f20c524912ee95f48e32a1d79629be6a033af3"} Nov 25 12:20:40 crc kubenswrapper[4715]: I1125 12:20:40.156446 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-lf54k" event={"ID":"f6028855-f32e-407c-a4d5-66e8caa6db64","Type":"ContainerStarted","Data":"04e5a60fa4cc17aa201190176758c7531a3b4552867b6127cfe6ff4c3ed21634"} Nov 25 12:20:42 crc kubenswrapper[4715]: I1125 12:20:42.189731 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-rwp55" event={"ID":"d193542e-7b7f-4ee2-99dc-6db50e82c1c2","Type":"ContainerStarted","Data":"a374be81f9e0d3231a45e5271f92e6c45cd3775438c3b4a997e727df3d6665e5"} Nov 25 12:20:42 crc kubenswrapper[4715]: I1125 12:20:42.191023 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:42 crc kubenswrapper[4715]: I1125 12:20:42.192964 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-lf54k" event={"ID":"f6028855-f32e-407c-a4d5-66e8caa6db64","Type":"ContainerStarted","Data":"bb2dd6938e368f234f87187676b0b1ec50e1b54b12fd44b0fe01b518ace3a494"} Nov 25 12:20:42 crc kubenswrapper[4715]: I1125 12:20:42.193140 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-lf54k" Nov 25 12:20:42 crc kubenswrapper[4715]: I1125 12:20:42.212980 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-rwp55" podStartSLOduration=2.087131637 podStartE2EDuration="5.212962232s" podCreationTimestamp="2025-11-25 12:20:37 +0000 UTC" firstStartedPulling="2025-11-25 12:20:37.876386204 +0000 UTC m=+688.383889225" lastFinishedPulling="2025-11-25 12:20:41.002216799 +0000 UTC m=+691.509719820" observedRunningTime="2025-11-25 12:20:42.206625439 +0000 UTC m=+692.714128460" watchObservedRunningTime="2025-11-25 12:20:42.212962232 +0000 UTC m=+692.720465253" Nov 25 12:20:42 crc kubenswrapper[4715]: I1125 12:20:42.227005 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-lf54k" podStartSLOduration=3.513360481 podStartE2EDuration="5.226988377s" podCreationTimestamp="2025-11-25 12:20:37 +0000 UTC" firstStartedPulling="2025-11-25 12:20:39.289515909 +0000 UTC m=+689.797018930" lastFinishedPulling="2025-11-25 12:20:41.003143805 +0000 UTC m=+691.510646826" observedRunningTime="2025-11-25 12:20:42.221812677 +0000 UTC m=+692.729315698" watchObservedRunningTime="2025-11-25 12:20:42.226988377 +0000 UTC m=+692.734491398" Nov 25 12:20:45 crc kubenswrapper[4715]: I1125 12:20:45.212273 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" event={"ID":"874e4bd7-5cb5-47f2-b905-fe2dc002546b","Type":"ContainerStarted","Data":"20699c77f427505ed6a8e9ab5e1a3fc28f12c685a98e6107d3e656594081e750"} Nov 25 12:20:45 crc kubenswrapper[4715]: I1125 12:20:45.212853 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:45 crc kubenswrapper[4715]: I1125 12:20:45.214989 4715 generic.go:334] "Generic (PLEG): container finished" podID="0191af85-1aa7-4c92-a45d-fc502d5dbf8c" containerID="1e59932c8c4cc3fa4abac5d43f39c1b888979837524abee5b70af4fdaf4f7e51" exitCode=0 Nov 25 12:20:45 crc kubenswrapper[4715]: I1125 12:20:45.215040 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerDied","Data":"1e59932c8c4cc3fa4abac5d43f39c1b888979837524abee5b70af4fdaf4f7e51"} Nov 25 12:20:45 crc kubenswrapper[4715]: I1125 12:20:45.229738 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" podStartSLOduration=1.216006206 podStartE2EDuration="8.229719489s" podCreationTimestamp="2025-11-25 12:20:37 +0000 UTC" firstStartedPulling="2025-11-25 12:20:37.972512138 +0000 UTC m=+688.480015159" lastFinishedPulling="2025-11-25 12:20:44.986225421 +0000 UTC m=+695.493728442" observedRunningTime="2025-11-25 12:20:45.227283759 +0000 UTC m=+695.734786790" watchObservedRunningTime="2025-11-25 12:20:45.229719489 +0000 UTC m=+695.737222520" Nov 25 12:20:46 crc kubenswrapper[4715]: I1125 12:20:46.226811 4715 generic.go:334] "Generic (PLEG): container finished" podID="0191af85-1aa7-4c92-a45d-fc502d5dbf8c" containerID="0a53aebf55b5ebd531b785c70fccc022233349157117e55cabaf3ea3b749a2d6" exitCode=0 Nov 25 12:20:46 crc kubenswrapper[4715]: I1125 12:20:46.226975 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerDied","Data":"0a53aebf55b5ebd531b785c70fccc022233349157117e55cabaf3ea3b749a2d6"} Nov 25 12:20:47 crc kubenswrapper[4715]: I1125 12:20:47.234101 4715 generic.go:334] "Generic (PLEG): container finished" podID="0191af85-1aa7-4c92-a45d-fc502d5dbf8c" containerID="59ee315262d68a785bdc92c32409ebd606c9d5e3e98c4eadf37f1c99f2bd40ef" exitCode=0 Nov 25 12:20:47 crc kubenswrapper[4715]: I1125 12:20:47.234140 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerDied","Data":"59ee315262d68a785bdc92c32409ebd606c9d5e3e98c4eadf37f1c99f2bd40ef"} Nov 25 12:20:47 crc kubenswrapper[4715]: I1125 12:20:47.560084 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-rwp55" Nov 25 12:20:48 crc kubenswrapper[4715]: I1125 12:20:48.246571 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerStarted","Data":"960a001b3dd7b53a147e125446fac228167179aa000a7869081918bedd2deec5"} Nov 25 12:20:48 crc kubenswrapper[4715]: I1125 12:20:48.246610 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerStarted","Data":"4f36794378ce7b9b03b5e5e4ca9a3c9410131f92a375a4c3c57e9811b4cd3073"} Nov 25 12:20:48 crc kubenswrapper[4715]: I1125 12:20:48.246619 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerStarted","Data":"b6122e37b4372d9bc7b2d54576b45a4819aa2822197391cafbb19224393d103b"} Nov 25 12:20:48 crc kubenswrapper[4715]: I1125 12:20:48.246627 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerStarted","Data":"cf113db520c76ca3c996aa2bf1e010f903b7e671f2a8727043dcf36aad74dd86"} Nov 25 12:20:48 crc kubenswrapper[4715]: I1125 12:20:48.246636 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerStarted","Data":"1690e2f4678a7c7b8d8ec0fc0612f9bdcbd5dea2a02e8ea2be922bf524290d89"} Nov 25 12:20:48 crc kubenswrapper[4715]: I1125 12:20:48.246646 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ddwdz" event={"ID":"0191af85-1aa7-4c92-a45d-fc502d5dbf8c","Type":"ContainerStarted","Data":"62246c84223edb7bef267dfb772306779c9d306dbafa1bac592882180b239806"} Nov 25 12:20:48 crc kubenswrapper[4715]: I1125 12:20:48.247519 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:48 crc kubenswrapper[4715]: I1125 12:20:48.272359 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ddwdz" podStartSLOduration=4.436501703 podStartE2EDuration="11.272346372s" podCreationTimestamp="2025-11-25 12:20:37 +0000 UTC" firstStartedPulling="2025-11-25 12:20:38.169808803 +0000 UTC m=+688.677311824" lastFinishedPulling="2025-11-25 12:20:45.005653452 +0000 UTC m=+695.513156493" observedRunningTime="2025-11-25 12:20:48.269336745 +0000 UTC m=+698.776839766" watchObservedRunningTime="2025-11-25 12:20:48.272346372 +0000 UTC m=+698.779849393" Nov 25 12:20:49 crc kubenswrapper[4715]: I1125 12:20:49.038501 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-lf54k" Nov 25 12:20:53 crc kubenswrapper[4715]: I1125 12:20:53.064138 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:53 crc kubenswrapper[4715]: I1125 12:20:53.130754 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:57 crc kubenswrapper[4715]: I1125 12:20:57.781824 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-8msp7" Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.066100 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ddwdz" Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.564139 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-bf2vp"] Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.564844 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bf2vp" Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.569362 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.569670 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-gxdwj" Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.570254 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.578539 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-bf2vp"] Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.664719 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm48k\" (UniqueName: \"kubernetes.io/projected/168f6030-3aed-43de-a709-5381f939ec0a-kube-api-access-zm48k\") pod \"mariadb-operator-index-bf2vp\" (UID: \"168f6030-3aed-43de-a709-5381f939ec0a\") " pod="openstack-operators/mariadb-operator-index-bf2vp" Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.766335 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm48k\" (UniqueName: \"kubernetes.io/projected/168f6030-3aed-43de-a709-5381f939ec0a-kube-api-access-zm48k\") pod \"mariadb-operator-index-bf2vp\" (UID: \"168f6030-3aed-43de-a709-5381f939ec0a\") " pod="openstack-operators/mariadb-operator-index-bf2vp" Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.790976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm48k\" (UniqueName: \"kubernetes.io/projected/168f6030-3aed-43de-a709-5381f939ec0a-kube-api-access-zm48k\") pod \"mariadb-operator-index-bf2vp\" (UID: \"168f6030-3aed-43de-a709-5381f939ec0a\") " pod="openstack-operators/mariadb-operator-index-bf2vp" Nov 25 12:20:58 crc kubenswrapper[4715]: I1125 12:20:58.881342 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bf2vp" Nov 25 12:20:59 crc kubenswrapper[4715]: I1125 12:20:59.062581 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-bf2vp"] Nov 25 12:20:59 crc kubenswrapper[4715]: W1125 12:20:59.068830 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod168f6030_3aed_43de_a709_5381f939ec0a.slice/crio-fbba987b9e2c91216227b5497468ca18018d0986cd04e12252a38128b1a44d1f WatchSource:0}: Error finding container fbba987b9e2c91216227b5497468ca18018d0986cd04e12252a38128b1a44d1f: Status 404 returned error can't find the container with id fbba987b9e2c91216227b5497468ca18018d0986cd04e12252a38128b1a44d1f Nov 25 12:20:59 crc kubenswrapper[4715]: I1125 12:20:59.319919 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bf2vp" event={"ID":"168f6030-3aed-43de-a709-5381f939ec0a","Type":"ContainerStarted","Data":"fbba987b9e2c91216227b5497468ca18018d0986cd04e12252a38128b1a44d1f"} Nov 25 12:21:01 crc kubenswrapper[4715]: I1125 12:21:01.337614 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bf2vp" event={"ID":"168f6030-3aed-43de-a709-5381f939ec0a","Type":"ContainerStarted","Data":"306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4"} Nov 25 12:21:01 crc kubenswrapper[4715]: I1125 12:21:01.353138 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-bf2vp" podStartSLOduration=1.959985299 podStartE2EDuration="3.353119233s" podCreationTimestamp="2025-11-25 12:20:58 +0000 UTC" firstStartedPulling="2025-11-25 12:20:59.070603452 +0000 UTC m=+709.578106473" lastFinishedPulling="2025-11-25 12:21:00.463737386 +0000 UTC m=+710.971240407" observedRunningTime="2025-11-25 12:21:01.351419684 +0000 UTC m=+711.858922695" watchObservedRunningTime="2025-11-25 12:21:01.353119233 +0000 UTC m=+711.860622254" Nov 25 12:21:03 crc kubenswrapper[4715]: I1125 12:21:03.954667 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-bf2vp"] Nov 25 12:21:03 crc kubenswrapper[4715]: I1125 12:21:03.955543 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-bf2vp" podUID="168f6030-3aed-43de-a709-5381f939ec0a" containerName="registry-server" containerID="cri-o://306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4" gracePeriod=2 Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.301595 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bf2vp" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.333931 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm48k\" (UniqueName: \"kubernetes.io/projected/168f6030-3aed-43de-a709-5381f939ec0a-kube-api-access-zm48k\") pod \"168f6030-3aed-43de-a709-5381f939ec0a\" (UID: \"168f6030-3aed-43de-a709-5381f939ec0a\") " Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.339329 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/168f6030-3aed-43de-a709-5381f939ec0a-kube-api-access-zm48k" (OuterVolumeSpecName: "kube-api-access-zm48k") pod "168f6030-3aed-43de-a709-5381f939ec0a" (UID: "168f6030-3aed-43de-a709-5381f939ec0a"). InnerVolumeSpecName "kube-api-access-zm48k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.357329 4715 generic.go:334] "Generic (PLEG): container finished" podID="168f6030-3aed-43de-a709-5381f939ec0a" containerID="306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4" exitCode=0 Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.357373 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bf2vp" event={"ID":"168f6030-3aed-43de-a709-5381f939ec0a","Type":"ContainerDied","Data":"306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4"} Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.357398 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bf2vp" event={"ID":"168f6030-3aed-43de-a709-5381f939ec0a","Type":"ContainerDied","Data":"fbba987b9e2c91216227b5497468ca18018d0986cd04e12252a38128b1a44d1f"} Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.357416 4715 scope.go:117] "RemoveContainer" containerID="306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.357598 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bf2vp" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.375311 4715 scope.go:117] "RemoveContainer" containerID="306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4" Nov 25 12:21:04 crc kubenswrapper[4715]: E1125 12:21:04.376591 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4\": container with ID starting with 306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4 not found: ID does not exist" containerID="306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.376709 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4"} err="failed to get container status \"306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4\": rpc error: code = NotFound desc = could not find container \"306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4\": container with ID starting with 306f9b421fab8a21c94117f3a8c3ad7e18719c01add2427f9281f6f7b046a8e4 not found: ID does not exist" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.386370 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-bf2vp"] Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.390739 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-bf2vp"] Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.435998 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm48k\" (UniqueName: \"kubernetes.io/projected/168f6030-3aed-43de-a709-5381f939ec0a-kube-api-access-zm48k\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.559672 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-wwd85"] Nov 25 12:21:04 crc kubenswrapper[4715]: E1125 12:21:04.559888 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="168f6030-3aed-43de-a709-5381f939ec0a" containerName="registry-server" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.559900 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="168f6030-3aed-43de-a709-5381f939ec0a" containerName="registry-server" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.560010 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="168f6030-3aed-43de-a709-5381f939ec0a" containerName="registry-server" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.560463 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-wwd85" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.561842 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-gxdwj" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.562681 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.568421 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-wwd85"] Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.568779 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.638949 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ghkq\" (UniqueName: \"kubernetes.io/projected/b470a68e-dbe6-4f0e-99fd-1c23ce4021bc-kube-api-access-6ghkq\") pod \"mariadb-operator-index-wwd85\" (UID: \"b470a68e-dbe6-4f0e-99fd-1c23ce4021bc\") " pod="openstack-operators/mariadb-operator-index-wwd85" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.701932 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="168f6030-3aed-43de-a709-5381f939ec0a" path="/var/lib/kubelet/pods/168f6030-3aed-43de-a709-5381f939ec0a/volumes" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.740913 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ghkq\" (UniqueName: \"kubernetes.io/projected/b470a68e-dbe6-4f0e-99fd-1c23ce4021bc-kube-api-access-6ghkq\") pod \"mariadb-operator-index-wwd85\" (UID: \"b470a68e-dbe6-4f0e-99fd-1c23ce4021bc\") " pod="openstack-operators/mariadb-operator-index-wwd85" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.757917 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ghkq\" (UniqueName: \"kubernetes.io/projected/b470a68e-dbe6-4f0e-99fd-1c23ce4021bc-kube-api-access-6ghkq\") pod \"mariadb-operator-index-wwd85\" (UID: \"b470a68e-dbe6-4f0e-99fd-1c23ce4021bc\") " pod="openstack-operators/mariadb-operator-index-wwd85" Nov 25 12:21:04 crc kubenswrapper[4715]: I1125 12:21:04.879722 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-wwd85" Nov 25 12:21:05 crc kubenswrapper[4715]: I1125 12:21:05.253598 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-wwd85"] Nov 25 12:21:05 crc kubenswrapper[4715]: I1125 12:21:05.364424 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-wwd85" event={"ID":"b470a68e-dbe6-4f0e-99fd-1c23ce4021bc","Type":"ContainerStarted","Data":"ff31a69f0a2efd7ad0dc1c386ba22cebb6bc324cf4c5191e73b3eaa5be4ec63d"} Nov 25 12:21:06 crc kubenswrapper[4715]: I1125 12:21:06.372061 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-wwd85" event={"ID":"b470a68e-dbe6-4f0e-99fd-1c23ce4021bc","Type":"ContainerStarted","Data":"e233b53456c029239a76b2d9968625ae28168b2e56d5b49a985529508d219e7f"} Nov 25 12:21:06 crc kubenswrapper[4715]: I1125 12:21:06.391809 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-wwd85" podStartSLOduration=1.6657508559999998 podStartE2EDuration="2.391784368s" podCreationTimestamp="2025-11-25 12:21:04 +0000 UTC" firstStartedPulling="2025-11-25 12:21:05.259012183 +0000 UTC m=+715.766515214" lastFinishedPulling="2025-11-25 12:21:05.985045695 +0000 UTC m=+716.492548726" observedRunningTime="2025-11-25 12:21:06.384833126 +0000 UTC m=+716.892336147" watchObservedRunningTime="2025-11-25 12:21:06.391784368 +0000 UTC m=+716.899287409" Nov 25 12:21:14 crc kubenswrapper[4715]: I1125 12:21:14.880598 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-wwd85" Nov 25 12:21:14 crc kubenswrapper[4715]: I1125 12:21:14.881286 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-wwd85" Nov 25 12:21:14 crc kubenswrapper[4715]: I1125 12:21:14.913103 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-wwd85" Nov 25 12:21:15 crc kubenswrapper[4715]: I1125 12:21:15.456531 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-wwd85" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.009406 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg"] Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.011139 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.013263 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7xsfb" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.022690 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg"] Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.105326 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-util\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.105697 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-bundle\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.105863 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttfnx\" (UniqueName: \"kubernetes.io/projected/68d8479f-beb5-4104-a9ca-a83b0c730f8a-kube-api-access-ttfnx\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.207221 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-util\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.207299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-bundle\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.207372 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttfnx\" (UniqueName: \"kubernetes.io/projected/68d8479f-beb5-4104-a9ca-a83b0c730f8a-kube-api-access-ttfnx\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.208019 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-bundle\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.208150 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-util\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.229895 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttfnx\" (UniqueName: \"kubernetes.io/projected/68d8479f-beb5-4104-a9ca-a83b0c730f8a-kube-api-access-ttfnx\") pod \"c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.338939 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:17 crc kubenswrapper[4715]: I1125 12:21:17.766603 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg"] Nov 25 12:21:18 crc kubenswrapper[4715]: I1125 12:21:18.451237 4715 generic.go:334] "Generic (PLEG): container finished" podID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerID="2fb5db05a276025b0de4568a6567f4c7a7c81be6f3f3b6484c6886450e2cd0d6" exitCode=0 Nov 25 12:21:18 crc kubenswrapper[4715]: I1125 12:21:18.451338 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" event={"ID":"68d8479f-beb5-4104-a9ca-a83b0c730f8a","Type":"ContainerDied","Data":"2fb5db05a276025b0de4568a6567f4c7a7c81be6f3f3b6484c6886450e2cd0d6"} Nov 25 12:21:18 crc kubenswrapper[4715]: I1125 12:21:18.451535 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" event={"ID":"68d8479f-beb5-4104-a9ca-a83b0c730f8a","Type":"ContainerStarted","Data":"40dc372689a68f1ab8b8d33adf53accb37b0ad272db0ee63c5f85976682b928a"} Nov 25 12:21:19 crc kubenswrapper[4715]: I1125 12:21:19.467736 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" event={"ID":"68d8479f-beb5-4104-a9ca-a83b0c730f8a","Type":"ContainerStarted","Data":"f6cee86d4f04f63fa3319d3ddf49d6b18a0b63f8c83d2f4203ba624019dfadfd"} Nov 25 12:21:20 crc kubenswrapper[4715]: I1125 12:21:20.476624 4715 generic.go:334] "Generic (PLEG): container finished" podID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerID="f6cee86d4f04f63fa3319d3ddf49d6b18a0b63f8c83d2f4203ba624019dfadfd" exitCode=0 Nov 25 12:21:20 crc kubenswrapper[4715]: I1125 12:21:20.476688 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" event={"ID":"68d8479f-beb5-4104-a9ca-a83b0c730f8a","Type":"ContainerDied","Data":"f6cee86d4f04f63fa3319d3ddf49d6b18a0b63f8c83d2f4203ba624019dfadfd"} Nov 25 12:21:21 crc kubenswrapper[4715]: I1125 12:21:21.484376 4715 generic.go:334] "Generic (PLEG): container finished" podID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerID="762e74dbcc616445764dd3814ed6f2d1a8339e38ac16a801850d03e919a34b7d" exitCode=0 Nov 25 12:21:21 crc kubenswrapper[4715]: I1125 12:21:21.484459 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" event={"ID":"68d8479f-beb5-4104-a9ca-a83b0c730f8a","Type":"ContainerDied","Data":"762e74dbcc616445764dd3814ed6f2d1a8339e38ac16a801850d03e919a34b7d"} Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.691375 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.787136 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-bundle\") pod \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.787363 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttfnx\" (UniqueName: \"kubernetes.io/projected/68d8479f-beb5-4104-a9ca-a83b0c730f8a-kube-api-access-ttfnx\") pod \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.787395 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-util\") pod \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\" (UID: \"68d8479f-beb5-4104-a9ca-a83b0c730f8a\") " Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.788463 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-bundle" (OuterVolumeSpecName: "bundle") pod "68d8479f-beb5-4104-a9ca-a83b0c730f8a" (UID: "68d8479f-beb5-4104-a9ca-a83b0c730f8a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.793431 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d8479f-beb5-4104-a9ca-a83b0c730f8a-kube-api-access-ttfnx" (OuterVolumeSpecName: "kube-api-access-ttfnx") pod "68d8479f-beb5-4104-a9ca-a83b0c730f8a" (UID: "68d8479f-beb5-4104-a9ca-a83b0c730f8a"). InnerVolumeSpecName "kube-api-access-ttfnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.807738 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-util" (OuterVolumeSpecName: "util") pod "68d8479f-beb5-4104-a9ca-a83b0c730f8a" (UID: "68d8479f-beb5-4104-a9ca-a83b0c730f8a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.888890 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.888933 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttfnx\" (UniqueName: \"kubernetes.io/projected/68d8479f-beb5-4104-a9ca-a83b0c730f8a-kube-api-access-ttfnx\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:22 crc kubenswrapper[4715]: I1125 12:21:22.888948 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/68d8479f-beb5-4104-a9ca-a83b0c730f8a-util\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:23 crc kubenswrapper[4715]: I1125 12:21:23.499375 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" event={"ID":"68d8479f-beb5-4104-a9ca-a83b0c730f8a","Type":"ContainerDied","Data":"40dc372689a68f1ab8b8d33adf53accb37b0ad272db0ee63c5f85976682b928a"} Nov 25 12:21:23 crc kubenswrapper[4715]: I1125 12:21:23.499417 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40dc372689a68f1ab8b8d33adf53accb37b0ad272db0ee63c5f85976682b928a" Nov 25 12:21:23 crc kubenswrapper[4715]: I1125 12:21:23.499431 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.349158 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp"] Nov 25 12:21:30 crc kubenswrapper[4715]: E1125 12:21:30.349955 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerName="extract" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.349971 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerName="extract" Nov 25 12:21:30 crc kubenswrapper[4715]: E1125 12:21:30.349983 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerName="util" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.349990 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerName="util" Nov 25 12:21:30 crc kubenswrapper[4715]: E1125 12:21:30.350001 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerName="pull" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.350007 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerName="pull" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.350132 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d8479f-beb5-4104-a9ca-a83b0c730f8a" containerName="extract" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.350942 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.352900 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.353655 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.357060 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qnsps" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.364510 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp"] Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.382854 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gf8s\" (UniqueName: \"kubernetes.io/projected/9fb80822-9010-4c95-8c77-225fc5a92b66-kube-api-access-7gf8s\") pod \"mariadb-operator-controller-manager-6b6d8bbdd9-78tmp\" (UID: \"9fb80822-9010-4c95-8c77-225fc5a92b66\") " pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.382923 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9fb80822-9010-4c95-8c77-225fc5a92b66-apiservice-cert\") pod \"mariadb-operator-controller-manager-6b6d8bbdd9-78tmp\" (UID: \"9fb80822-9010-4c95-8c77-225fc5a92b66\") " pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.383056 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9fb80822-9010-4c95-8c77-225fc5a92b66-webhook-cert\") pod \"mariadb-operator-controller-manager-6b6d8bbdd9-78tmp\" (UID: \"9fb80822-9010-4c95-8c77-225fc5a92b66\") " pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.484446 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9fb80822-9010-4c95-8c77-225fc5a92b66-webhook-cert\") pod \"mariadb-operator-controller-manager-6b6d8bbdd9-78tmp\" (UID: \"9fb80822-9010-4c95-8c77-225fc5a92b66\") " pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.484507 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gf8s\" (UniqueName: \"kubernetes.io/projected/9fb80822-9010-4c95-8c77-225fc5a92b66-kube-api-access-7gf8s\") pod \"mariadb-operator-controller-manager-6b6d8bbdd9-78tmp\" (UID: \"9fb80822-9010-4c95-8c77-225fc5a92b66\") " pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.484546 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9fb80822-9010-4c95-8c77-225fc5a92b66-apiservice-cert\") pod \"mariadb-operator-controller-manager-6b6d8bbdd9-78tmp\" (UID: \"9fb80822-9010-4c95-8c77-225fc5a92b66\") " pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.489875 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9fb80822-9010-4c95-8c77-225fc5a92b66-webhook-cert\") pod \"mariadb-operator-controller-manager-6b6d8bbdd9-78tmp\" (UID: \"9fb80822-9010-4c95-8c77-225fc5a92b66\") " pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.489901 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9fb80822-9010-4c95-8c77-225fc5a92b66-apiservice-cert\") pod \"mariadb-operator-controller-manager-6b6d8bbdd9-78tmp\" (UID: \"9fb80822-9010-4c95-8c77-225fc5a92b66\") " pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.501564 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gf8s\" (UniqueName: \"kubernetes.io/projected/9fb80822-9010-4c95-8c77-225fc5a92b66-kube-api-access-7gf8s\") pod \"mariadb-operator-controller-manager-6b6d8bbdd9-78tmp\" (UID: \"9fb80822-9010-4c95-8c77-225fc5a92b66\") " pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.671580 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:30 crc kubenswrapper[4715]: I1125 12:21:30.881771 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp"] Nov 25 12:21:31 crc kubenswrapper[4715]: I1125 12:21:31.546785 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" event={"ID":"9fb80822-9010-4c95-8c77-225fc5a92b66","Type":"ContainerStarted","Data":"981780ebeca4199214019f605f37e71f4129a45905b79cddb819d71535906856"} Nov 25 12:21:34 crc kubenswrapper[4715]: I1125 12:21:34.563038 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" event={"ID":"9fb80822-9010-4c95-8c77-225fc5a92b66","Type":"ContainerStarted","Data":"9330f754e7a93df97302538ab691b89d3ce5cdcd67c19332f3c5b1967b63d148"} Nov 25 12:21:34 crc kubenswrapper[4715]: I1125 12:21:34.563500 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:34 crc kubenswrapper[4715]: I1125 12:21:34.580877 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" podStartSLOduration=1.059134502 podStartE2EDuration="4.580856549s" podCreationTimestamp="2025-11-25 12:21:30 +0000 UTC" firstStartedPulling="2025-11-25 12:21:30.890331703 +0000 UTC m=+741.397834724" lastFinishedPulling="2025-11-25 12:21:34.41205375 +0000 UTC m=+744.919556771" observedRunningTime="2025-11-25 12:21:34.575537092 +0000 UTC m=+745.083040113" watchObservedRunningTime="2025-11-25 12:21:34.580856549 +0000 UTC m=+745.088359580" Nov 25 12:21:40 crc kubenswrapper[4715]: I1125 12:21:40.679339 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6b6d8bbdd9-78tmp" Nov 25 12:21:43 crc kubenswrapper[4715]: I1125 12:21:43.705215 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-2mp82"] Nov 25 12:21:43 crc kubenswrapper[4715]: I1125 12:21:43.706302 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-2mp82" Nov 25 12:21:43 crc kubenswrapper[4715]: I1125 12:21:43.717929 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-mtt9q" Nov 25 12:21:43 crc kubenswrapper[4715]: I1125 12:21:43.724026 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-2mp82"] Nov 25 12:21:43 crc kubenswrapper[4715]: I1125 12:21:43.760881 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl7f2\" (UniqueName: \"kubernetes.io/projected/8f728d84-3952-4b88-9fb9-565cfd25abc6-kube-api-access-tl7f2\") pod \"infra-operator-index-2mp82\" (UID: \"8f728d84-3952-4b88-9fb9-565cfd25abc6\") " pod="openstack-operators/infra-operator-index-2mp82" Nov 25 12:21:43 crc kubenswrapper[4715]: I1125 12:21:43.862211 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl7f2\" (UniqueName: \"kubernetes.io/projected/8f728d84-3952-4b88-9fb9-565cfd25abc6-kube-api-access-tl7f2\") pod \"infra-operator-index-2mp82\" (UID: \"8f728d84-3952-4b88-9fb9-565cfd25abc6\") " pod="openstack-operators/infra-operator-index-2mp82" Nov 25 12:21:43 crc kubenswrapper[4715]: I1125 12:21:43.879432 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl7f2\" (UniqueName: \"kubernetes.io/projected/8f728d84-3952-4b88-9fb9-565cfd25abc6-kube-api-access-tl7f2\") pod \"infra-operator-index-2mp82\" (UID: \"8f728d84-3952-4b88-9fb9-565cfd25abc6\") " pod="openstack-operators/infra-operator-index-2mp82" Nov 25 12:21:44 crc kubenswrapper[4715]: I1125 12:21:44.020163 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-2mp82" Nov 25 12:21:44 crc kubenswrapper[4715]: I1125 12:21:44.443008 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-2mp82"] Nov 25 12:21:44 crc kubenswrapper[4715]: I1125 12:21:44.626869 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-2mp82" event={"ID":"8f728d84-3952-4b88-9fb9-565cfd25abc6","Type":"ContainerStarted","Data":"863eae0bb3fcf6fdb0b54c551dd02767d7ee6482de4a65f1bd551dd0058cb591"} Nov 25 12:21:46 crc kubenswrapper[4715]: I1125 12:21:46.640835 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-2mp82" event={"ID":"8f728d84-3952-4b88-9fb9-565cfd25abc6","Type":"ContainerStarted","Data":"559b5e77dddea35834cc0442cd6ca26f72d25defaf9fe575e1e28f889a68bb11"} Nov 25 12:21:46 crc kubenswrapper[4715]: I1125 12:21:46.660585 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-2mp82" podStartSLOduration=1.9390871729999999 podStartE2EDuration="3.660570125s" podCreationTimestamp="2025-11-25 12:21:43 +0000 UTC" firstStartedPulling="2025-11-25 12:21:44.449519964 +0000 UTC m=+754.957022985" lastFinishedPulling="2025-11-25 12:21:46.171002876 +0000 UTC m=+756.678505937" observedRunningTime="2025-11-25 12:21:46.658028494 +0000 UTC m=+757.165531515" watchObservedRunningTime="2025-11-25 12:21:46.660570125 +0000 UTC m=+757.168073146" Nov 25 12:21:54 crc kubenswrapper[4715]: I1125 12:21:54.020890 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-2mp82" Nov 25 12:21:54 crc kubenswrapper[4715]: I1125 12:21:54.021531 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-2mp82" Nov 25 12:21:54 crc kubenswrapper[4715]: I1125 12:21:54.055427 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-2mp82" Nov 25 12:21:54 crc kubenswrapper[4715]: I1125 12:21:54.732403 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-2mp82" Nov 25 12:21:56 crc kubenswrapper[4715]: I1125 12:21:56.282054 4715 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 12:21:56 crc kubenswrapper[4715]: I1125 12:21:56.937774 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp"] Nov 25 12:21:56 crc kubenswrapper[4715]: I1125 12:21:56.939520 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:56 crc kubenswrapper[4715]: I1125 12:21:56.943452 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7xsfb" Nov 25 12:21:56 crc kubenswrapper[4715]: I1125 12:21:56.950040 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp"] Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.031678 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.032001 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.032127 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m29b8\" (UniqueName: \"kubernetes.io/projected/07d69d4e-e804-480e-86b1-1e0987585af8-kube-api-access-m29b8\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.133422 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.133519 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m29b8\" (UniqueName: \"kubernetes.io/projected/07d69d4e-e804-480e-86b1-1e0987585af8-kube-api-access-m29b8\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.133628 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.134216 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.134485 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.160914 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m29b8\" (UniqueName: \"kubernetes.io/projected/07d69d4e-e804-480e-86b1-1e0987585af8-kube-api-access-m29b8\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.256956 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.444116 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp"] Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.725858 4715 generic.go:334] "Generic (PLEG): container finished" podID="07d69d4e-e804-480e-86b1-1e0987585af8" containerID="151f6a44d0d0de438bd7a3a1b204ab95c8c2355a501c84e3cbd400dc1af03e64" exitCode=0 Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.725967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" event={"ID":"07d69d4e-e804-480e-86b1-1e0987585af8","Type":"ContainerDied","Data":"151f6a44d0d0de438bd7a3a1b204ab95c8c2355a501c84e3cbd400dc1af03e64"} Nov 25 12:21:57 crc kubenswrapper[4715]: I1125 12:21:57.726316 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" event={"ID":"07d69d4e-e804-480e-86b1-1e0987585af8","Type":"ContainerStarted","Data":"a637c8d78fc1155b7e02bb54dcf8bfaf88cbd2073aeffb4807cb44ee414e1406"} Nov 25 12:21:58 crc kubenswrapper[4715]: I1125 12:21:58.734576 4715 generic.go:334] "Generic (PLEG): container finished" podID="07d69d4e-e804-480e-86b1-1e0987585af8" containerID="6ca4bb021394d762d5af54adffda0cb92bdcf881e2751a3e752e52251dd862ca" exitCode=0 Nov 25 12:21:58 crc kubenswrapper[4715]: I1125 12:21:58.734622 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" event={"ID":"07d69d4e-e804-480e-86b1-1e0987585af8","Type":"ContainerDied","Data":"6ca4bb021394d762d5af54adffda0cb92bdcf881e2751a3e752e52251dd862ca"} Nov 25 12:21:59 crc kubenswrapper[4715]: I1125 12:21:59.745263 4715 generic.go:334] "Generic (PLEG): container finished" podID="07d69d4e-e804-480e-86b1-1e0987585af8" containerID="4f0d4f46cc88a2101fb70131164f858333449be7c5bad500d2b2d4ead568f8a2" exitCode=0 Nov 25 12:21:59 crc kubenswrapper[4715]: I1125 12:21:59.745353 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" event={"ID":"07d69d4e-e804-480e-86b1-1e0987585af8","Type":"ContainerDied","Data":"4f0d4f46cc88a2101fb70131164f858333449be7c5bad500d2b2d4ead568f8a2"} Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.001109 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.082870 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m29b8\" (UniqueName: \"kubernetes.io/projected/07d69d4e-e804-480e-86b1-1e0987585af8-kube-api-access-m29b8\") pod \"07d69d4e-e804-480e-86b1-1e0987585af8\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.082961 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-util\") pod \"07d69d4e-e804-480e-86b1-1e0987585af8\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.083016 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-bundle\") pod \"07d69d4e-e804-480e-86b1-1e0987585af8\" (UID: \"07d69d4e-e804-480e-86b1-1e0987585af8\") " Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.084218 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-bundle" (OuterVolumeSpecName: "bundle") pod "07d69d4e-e804-480e-86b1-1e0987585af8" (UID: "07d69d4e-e804-480e-86b1-1e0987585af8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.088890 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07d69d4e-e804-480e-86b1-1e0987585af8-kube-api-access-m29b8" (OuterVolumeSpecName: "kube-api-access-m29b8") pod "07d69d4e-e804-480e-86b1-1e0987585af8" (UID: "07d69d4e-e804-480e-86b1-1e0987585af8"). InnerVolumeSpecName "kube-api-access-m29b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.098464 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-util" (OuterVolumeSpecName: "util") pod "07d69d4e-e804-480e-86b1-1e0987585af8" (UID: "07d69d4e-e804-480e-86b1-1e0987585af8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.184107 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m29b8\" (UniqueName: \"kubernetes.io/projected/07d69d4e-e804-480e-86b1-1e0987585af8-kube-api-access-m29b8\") on node \"crc\" DevicePath \"\"" Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.184141 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-util\") on node \"crc\" DevicePath \"\"" Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.184151 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/07d69d4e-e804-480e-86b1-1e0987585af8-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.760033 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" event={"ID":"07d69d4e-e804-480e-86b1-1e0987585af8","Type":"ContainerDied","Data":"a637c8d78fc1155b7e02bb54dcf8bfaf88cbd2073aeffb4807cb44ee414e1406"} Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.760389 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a637c8d78fc1155b7e02bb54dcf8bfaf88cbd2073aeffb4807cb44ee414e1406" Nov 25 12:22:01 crc kubenswrapper[4715]: I1125 12:22:01.760118 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.171041 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl"] Nov 25 12:22:05 crc kubenswrapper[4715]: E1125 12:22:05.171525 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d69d4e-e804-480e-86b1-1e0987585af8" containerName="extract" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.171537 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d69d4e-e804-480e-86b1-1e0987585af8" containerName="extract" Nov 25 12:22:05 crc kubenswrapper[4715]: E1125 12:22:05.171545 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d69d4e-e804-480e-86b1-1e0987585af8" containerName="pull" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.171551 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d69d4e-e804-480e-86b1-1e0987585af8" containerName="pull" Nov 25 12:22:05 crc kubenswrapper[4715]: E1125 12:22:05.171562 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d69d4e-e804-480e-86b1-1e0987585af8" containerName="util" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.171568 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d69d4e-e804-480e-86b1-1e0987585af8" containerName="util" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.171660 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="07d69d4e-e804-480e-86b1-1e0987585af8" containerName="extract" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.172274 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.177157 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gthlk" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.177356 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.236744 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2dpr\" (UniqueName: \"kubernetes.io/projected/2f3c70c2-5a97-479c-aed9-151ba8a3c9d8-kube-api-access-b2dpr\") pod \"infra-operator-controller-manager-776cc6d755-b7wjl\" (UID: \"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8\") " pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.236817 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f3c70c2-5a97-479c-aed9-151ba8a3c9d8-webhook-cert\") pod \"infra-operator-controller-manager-776cc6d755-b7wjl\" (UID: \"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8\") " pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.236922 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f3c70c2-5a97-479c-aed9-151ba8a3c9d8-apiservice-cert\") pod \"infra-operator-controller-manager-776cc6d755-b7wjl\" (UID: \"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8\") " pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.240793 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl"] Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.337827 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f3c70c2-5a97-479c-aed9-151ba8a3c9d8-apiservice-cert\") pod \"infra-operator-controller-manager-776cc6d755-b7wjl\" (UID: \"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8\") " pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.337894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2dpr\" (UniqueName: \"kubernetes.io/projected/2f3c70c2-5a97-479c-aed9-151ba8a3c9d8-kube-api-access-b2dpr\") pod \"infra-operator-controller-manager-776cc6d755-b7wjl\" (UID: \"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8\") " pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.337925 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f3c70c2-5a97-479c-aed9-151ba8a3c9d8-webhook-cert\") pod \"infra-operator-controller-manager-776cc6d755-b7wjl\" (UID: \"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8\") " pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.343311 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f3c70c2-5a97-479c-aed9-151ba8a3c9d8-webhook-cert\") pod \"infra-operator-controller-manager-776cc6d755-b7wjl\" (UID: \"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8\") " pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.349211 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f3c70c2-5a97-479c-aed9-151ba8a3c9d8-apiservice-cert\") pod \"infra-operator-controller-manager-776cc6d755-b7wjl\" (UID: \"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8\") " pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.373668 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2dpr\" (UniqueName: \"kubernetes.io/projected/2f3c70c2-5a97-479c-aed9-151ba8a3c9d8-kube-api-access-b2dpr\") pod \"infra-operator-controller-manager-776cc6d755-b7wjl\" (UID: \"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8\") " pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.488711 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:05 crc kubenswrapper[4715]: I1125 12:22:05.915820 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl"] Nov 25 12:22:05 crc kubenswrapper[4715]: W1125 12:22:05.921264 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f3c70c2_5a97_479c_aed9_151ba8a3c9d8.slice/crio-378d415c9674d3b90792b36fcb32cf0eed5fbe813d0737e903378c56bcb169df WatchSource:0}: Error finding container 378d415c9674d3b90792b36fcb32cf0eed5fbe813d0737e903378c56bcb169df: Status 404 returned error can't find the container with id 378d415c9674d3b90792b36fcb32cf0eed5fbe813d0737e903378c56bcb169df Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.279812 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.280716 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.282810 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.283050 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.283306 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.283626 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-bp6pg" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.283796 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.292207 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.293591 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.297491 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.303057 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.304271 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.310908 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.317110 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.351830 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.351893 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/473c10ab-6523-402f-b66c-5d846a033294-operator-scripts\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.351923 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85183f2d-ee39-42f1-9153-9a7e6599ef2b-kolla-config\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.351949 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/473c10ab-6523-402f-b66c-5d846a033294-config-data-default\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.351973 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/85183f2d-ee39-42f1-9153-9a7e6599ef2b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.351993 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/85183f2d-ee39-42f1-9153-9a7e6599ef2b-config-data-default\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352046 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85n58\" (UniqueName: \"kubernetes.io/projected/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-kube-api-access-85n58\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352070 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/473c10ab-6523-402f-b66c-5d846a033294-config-data-generated\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352132 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85183f2d-ee39-42f1-9153-9a7e6599ef2b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352163 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-config-data-generated\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352208 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352241 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/473c10ab-6523-402f-b66c-5d846a033294-kolla-config\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352261 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv9rs\" (UniqueName: \"kubernetes.io/projected/473c10ab-6523-402f-b66c-5d846a033294-kube-api-access-nv9rs\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352286 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-operator-scripts\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352328 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv6ql\" (UniqueName: \"kubernetes.io/projected/85183f2d-ee39-42f1-9153-9a7e6599ef2b-kube-api-access-jv6ql\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352353 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-kolla-config\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352382 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.352416 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-config-data-default\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453492 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453541 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/473c10ab-6523-402f-b66c-5d846a033294-operator-scripts\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453575 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85183f2d-ee39-42f1-9153-9a7e6599ef2b-kolla-config\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453591 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/473c10ab-6523-402f-b66c-5d846a033294-config-data-default\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453606 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/85183f2d-ee39-42f1-9153-9a7e6599ef2b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453624 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/85183f2d-ee39-42f1-9153-9a7e6599ef2b-config-data-default\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453651 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85n58\" (UniqueName: \"kubernetes.io/projected/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-kube-api-access-85n58\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453674 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/473c10ab-6523-402f-b66c-5d846a033294-config-data-generated\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453701 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85183f2d-ee39-42f1-9153-9a7e6599ef2b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453728 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-config-data-generated\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453745 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453766 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/473c10ab-6523-402f-b66c-5d846a033294-kolla-config\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453782 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv9rs\" (UniqueName: \"kubernetes.io/projected/473c10ab-6523-402f-b66c-5d846a033294-kube-api-access-nv9rs\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453799 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-operator-scripts\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453830 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv6ql\" (UniqueName: \"kubernetes.io/projected/85183f2d-ee39-42f1-9153-9a7e6599ef2b-kube-api-access-jv6ql\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453849 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-kolla-config\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453871 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-config-data-default\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.453889 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.454278 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.454301 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.454874 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-kolla-config\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.455065 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.455419 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/85183f2d-ee39-42f1-9153-9a7e6599ef2b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.455555 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85183f2d-ee39-42f1-9153-9a7e6599ef2b-kolla-config\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.455926 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-config-data-generated\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.456420 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/85183f2d-ee39-42f1-9153-9a7e6599ef2b-config-data-default\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.456518 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85183f2d-ee39-42f1-9153-9a7e6599ef2b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.456648 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-operator-scripts\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.456816 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/473c10ab-6523-402f-b66c-5d846a033294-config-data-generated\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.457240 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/473c10ab-6523-402f-b66c-5d846a033294-kolla-config\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.457315 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/473c10ab-6523-402f-b66c-5d846a033294-config-data-default\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.457964 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/473c10ab-6523-402f-b66c-5d846a033294-operator-scripts\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.460537 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-config-data-default\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.473632 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.474177 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.474574 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.475894 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv9rs\" (UniqueName: \"kubernetes.io/projected/473c10ab-6523-402f-b66c-5d846a033294-kube-api-access-nv9rs\") pod \"openstack-galera-2\" (UID: \"473c10ab-6523-402f-b66c-5d846a033294\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.479660 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85n58\" (UniqueName: \"kubernetes.io/projected/8fb56a14-d3f7-4cf3-ae0b-8aad521a3423-kube-api-access-85n58\") pod \"openstack-galera-1\" (UID: \"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.480528 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv6ql\" (UniqueName: \"kubernetes.io/projected/85183f2d-ee39-42f1-9153-9a7e6599ef2b-kube-api-access-jv6ql\") pod \"openstack-galera-0\" (UID: \"85183f2d-ee39-42f1-9153-9a7e6599ef2b\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.599451 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.633042 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.638291 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.801203 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" event={"ID":"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8","Type":"ContainerStarted","Data":"378d415c9674d3b90792b36fcb32cf0eed5fbe813d0737e903378c56bcb169df"} Nov 25 12:22:06 crc kubenswrapper[4715]: I1125 12:22:06.956624 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 25 12:22:06 crc kubenswrapper[4715]: W1125 12:22:06.960582 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fb56a14_d3f7_4cf3_ae0b_8aad521a3423.slice/crio-aa9f7ce1b72535a42b4e2bf726d66dca02981c32565ef30bb5b2658c5e3ee5c7 WatchSource:0}: Error finding container aa9f7ce1b72535a42b4e2bf726d66dca02981c32565ef30bb5b2658c5e3ee5c7: Status 404 returned error can't find the container with id aa9f7ce1b72535a42b4e2bf726d66dca02981c32565ef30bb5b2658c5e3ee5c7 Nov 25 12:22:07 crc kubenswrapper[4715]: I1125 12:22:07.027450 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 25 12:22:07 crc kubenswrapper[4715]: W1125 12:22:07.032036 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85183f2d_ee39_42f1_9153_9a7e6599ef2b.slice/crio-132f02f44e8377c1e07ebf83351212df9c81023c9e75c9da853e5f925d03562d WatchSource:0}: Error finding container 132f02f44e8377c1e07ebf83351212df9c81023c9e75c9da853e5f925d03562d: Status 404 returned error can't find the container with id 132f02f44e8377c1e07ebf83351212df9c81023c9e75c9da853e5f925d03562d Nov 25 12:22:07 crc kubenswrapper[4715]: I1125 12:22:07.112946 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 25 12:22:07 crc kubenswrapper[4715]: W1125 12:22:07.119929 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod473c10ab_6523_402f_b66c_5d846a033294.slice/crio-0904f4475c7903e5e6c381ba910d045521d3785cecb3d0ae85f7a2b23b667a78 WatchSource:0}: Error finding container 0904f4475c7903e5e6c381ba910d045521d3785cecb3d0ae85f7a2b23b667a78: Status 404 returned error can't find the container with id 0904f4475c7903e5e6c381ba910d045521d3785cecb3d0ae85f7a2b23b667a78 Nov 25 12:22:07 crc kubenswrapper[4715]: I1125 12:22:07.809504 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"473c10ab-6523-402f-b66c-5d846a033294","Type":"ContainerStarted","Data":"0904f4475c7903e5e6c381ba910d045521d3785cecb3d0ae85f7a2b23b667a78"} Nov 25 12:22:07 crc kubenswrapper[4715]: I1125 12:22:07.810910 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423","Type":"ContainerStarted","Data":"aa9f7ce1b72535a42b4e2bf726d66dca02981c32565ef30bb5b2658c5e3ee5c7"} Nov 25 12:22:07 crc kubenswrapper[4715]: I1125 12:22:07.811693 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"85183f2d-ee39-42f1-9153-9a7e6599ef2b","Type":"ContainerStarted","Data":"132f02f44e8377c1e07ebf83351212df9c81023c9e75c9da853e5f925d03562d"} Nov 25 12:22:08 crc kubenswrapper[4715]: I1125 12:22:08.823565 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" event={"ID":"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8","Type":"ContainerStarted","Data":"bf524f7bd49954a16f4139ebb9cb716e2b8e77cd0fc7329160e209b9d731209f"} Nov 25 12:22:15 crc kubenswrapper[4715]: I1125 12:22:15.876406 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"473c10ab-6523-402f-b66c-5d846a033294","Type":"ContainerStarted","Data":"c8a78175fc288fe61c57c0988533f0a101198a8a5d2f3acb2f301910c172aaa4"} Nov 25 12:22:15 crc kubenswrapper[4715]: I1125 12:22:15.880310 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" event={"ID":"2f3c70c2-5a97-479c-aed9-151ba8a3c9d8","Type":"ContainerStarted","Data":"3470e6eb73c92bc8acc0c0fcea409636f392b96b807dc8f06fad057f51556240"} Nov 25 12:22:15 crc kubenswrapper[4715]: I1125 12:22:15.880539 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:15 crc kubenswrapper[4715]: I1125 12:22:15.881971 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423","Type":"ContainerStarted","Data":"d3e6bd17f62a17dae5d9a2d53434268d31ec526d42034827564c4c6cf6f689cf"} Nov 25 12:22:15 crc kubenswrapper[4715]: I1125 12:22:15.883612 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"85183f2d-ee39-42f1-9153-9a7e6599ef2b","Type":"ContainerStarted","Data":"eb9baf1e14a2dabdaaed4c874c42948af0dc838abbbd2271aafd54b1026809d1"} Nov 25 12:22:15 crc kubenswrapper[4715]: I1125 12:22:15.885075 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" Nov 25 12:22:15 crc kubenswrapper[4715]: I1125 12:22:15.943450 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-776cc6d755-b7wjl" podStartSLOduration=1.5506230269999999 podStartE2EDuration="10.943423888s" podCreationTimestamp="2025-11-25 12:22:05 +0000 UTC" firstStartedPulling="2025-11-25 12:22:05.923249982 +0000 UTC m=+776.430753003" lastFinishedPulling="2025-11-25 12:22:15.316050823 +0000 UTC m=+785.823553864" observedRunningTime="2025-11-25 12:22:15.924489827 +0000 UTC m=+786.431992848" watchObservedRunningTime="2025-11-25 12:22:15.943423888 +0000 UTC m=+786.450926909" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.602797 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.603529 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.612496 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-7p9rn" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.612778 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.621109 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.707498 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f48ce5a-f968-4dc7-a0bb-eff579f590af-config-data\") pod \"memcached-0\" (UID: \"7f48ce5a-f968-4dc7-a0bb-eff579f590af\") " pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.707555 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7f48ce5a-f968-4dc7-a0bb-eff579f590af-kolla-config\") pod \"memcached-0\" (UID: \"7f48ce5a-f968-4dc7-a0bb-eff579f590af\") " pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.707717 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsj95\" (UniqueName: \"kubernetes.io/projected/7f48ce5a-f968-4dc7-a0bb-eff579f590af-kube-api-access-bsj95\") pod \"memcached-0\" (UID: \"7f48ce5a-f968-4dc7-a0bb-eff579f590af\") " pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.809321 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f48ce5a-f968-4dc7-a0bb-eff579f590af-config-data\") pod \"memcached-0\" (UID: \"7f48ce5a-f968-4dc7-a0bb-eff579f590af\") " pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.809418 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7f48ce5a-f968-4dc7-a0bb-eff579f590af-kolla-config\") pod \"memcached-0\" (UID: \"7f48ce5a-f968-4dc7-a0bb-eff579f590af\") " pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.809534 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsj95\" (UniqueName: \"kubernetes.io/projected/7f48ce5a-f968-4dc7-a0bb-eff579f590af-kube-api-access-bsj95\") pod \"memcached-0\" (UID: \"7f48ce5a-f968-4dc7-a0bb-eff579f590af\") " pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.811869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f48ce5a-f968-4dc7-a0bb-eff579f590af-config-data\") pod \"memcached-0\" (UID: \"7f48ce5a-f968-4dc7-a0bb-eff579f590af\") " pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.812740 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7f48ce5a-f968-4dc7-a0bb-eff579f590af-kolla-config\") pod \"memcached-0\" (UID: \"7f48ce5a-f968-4dc7-a0bb-eff579f590af\") " pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.835088 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsj95\" (UniqueName: \"kubernetes.io/projected/7f48ce5a-f968-4dc7-a0bb-eff579f590af-kube-api-access-bsj95\") pod \"memcached-0\" (UID: \"7f48ce5a-f968-4dc7-a0bb-eff579f590af\") " pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:16 crc kubenswrapper[4715]: I1125 12:22:16.921023 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:17 crc kubenswrapper[4715]: I1125 12:22:17.331431 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 25 12:22:17 crc kubenswrapper[4715]: W1125 12:22:17.339485 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f48ce5a_f968_4dc7_a0bb_eff579f590af.slice/crio-e82d274c1113d28131d6d8ba2c5bf5f6650e485b71569ee613475bc49937dce9 WatchSource:0}: Error finding container e82d274c1113d28131d6d8ba2c5bf5f6650e485b71569ee613475bc49937dce9: Status 404 returned error can't find the container with id e82d274c1113d28131d6d8ba2c5bf5f6650e485b71569ee613475bc49937dce9 Nov 25 12:22:17 crc kubenswrapper[4715]: I1125 12:22:17.894738 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"7f48ce5a-f968-4dc7-a0bb-eff579f590af","Type":"ContainerStarted","Data":"e82d274c1113d28131d6d8ba2c5bf5f6650e485b71569ee613475bc49937dce9"} Nov 25 12:22:17 crc kubenswrapper[4715]: I1125 12:22:17.943927 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:22:17 crc kubenswrapper[4715]: I1125 12:22:17.944040 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.512985 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-cf7p9"] Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.514659 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.517550 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-hbw9h" Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.522252 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-cf7p9"] Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.661040 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96x7h\" (UniqueName: \"kubernetes.io/projected/b1c71c09-df63-430f-80f2-74d6cdb35e1a-kube-api-access-96x7h\") pod \"rabbitmq-cluster-operator-index-cf7p9\" (UID: \"b1c71c09-df63-430f-80f2-74d6cdb35e1a\") " pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.763624 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96x7h\" (UniqueName: \"kubernetes.io/projected/b1c71c09-df63-430f-80f2-74d6cdb35e1a-kube-api-access-96x7h\") pod \"rabbitmq-cluster-operator-index-cf7p9\" (UID: \"b1c71c09-df63-430f-80f2-74d6cdb35e1a\") " pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.794998 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96x7h\" (UniqueName: \"kubernetes.io/projected/b1c71c09-df63-430f-80f2-74d6cdb35e1a-kube-api-access-96x7h\") pod \"rabbitmq-cluster-operator-index-cf7p9\" (UID: \"b1c71c09-df63-430f-80f2-74d6cdb35e1a\") " pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.846063 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.911092 4715 generic.go:334] "Generic (PLEG): container finished" podID="8fb56a14-d3f7-4cf3-ae0b-8aad521a3423" containerID="d3e6bd17f62a17dae5d9a2d53434268d31ec526d42034827564c4c6cf6f689cf" exitCode=0 Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.911441 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423","Type":"ContainerDied","Data":"d3e6bd17f62a17dae5d9a2d53434268d31ec526d42034827564c4c6cf6f689cf"} Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.914353 4715 generic.go:334] "Generic (PLEG): container finished" podID="85183f2d-ee39-42f1-9153-9a7e6599ef2b" containerID="eb9baf1e14a2dabdaaed4c874c42948af0dc838abbbd2271aafd54b1026809d1" exitCode=0 Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.914413 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"85183f2d-ee39-42f1-9153-9a7e6599ef2b","Type":"ContainerDied","Data":"eb9baf1e14a2dabdaaed4c874c42948af0dc838abbbd2271aafd54b1026809d1"} Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.916839 4715 generic.go:334] "Generic (PLEG): container finished" podID="473c10ab-6523-402f-b66c-5d846a033294" containerID="c8a78175fc288fe61c57c0988533f0a101198a8a5d2f3acb2f301910c172aaa4" exitCode=0 Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.916905 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"473c10ab-6523-402f-b66c-5d846a033294","Type":"ContainerDied","Data":"c8a78175fc288fe61c57c0988533f0a101198a8a5d2f3acb2f301910c172aaa4"} Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.924855 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"7f48ce5a-f968-4dc7-a0bb-eff579f590af","Type":"ContainerStarted","Data":"7925ef34701d1060456c0f53a0e99181f93507f00f9c18cb51cb8c25fcdef8a4"} Nov 25 12:22:19 crc kubenswrapper[4715]: I1125 12:22:19.925828 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:20 crc kubenswrapper[4715]: I1125 12:22:20.001166 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=2.114436708 podStartE2EDuration="4.00114171s" podCreationTimestamp="2025-11-25 12:22:16 +0000 UTC" firstStartedPulling="2025-11-25 12:22:17.341331907 +0000 UTC m=+787.848834928" lastFinishedPulling="2025-11-25 12:22:19.228036919 +0000 UTC m=+789.735539930" observedRunningTime="2025-11-25 12:22:19.98569803 +0000 UTC m=+790.493201051" watchObservedRunningTime="2025-11-25 12:22:20.00114171 +0000 UTC m=+790.508644731" Nov 25 12:22:20 crc kubenswrapper[4715]: I1125 12:22:20.163371 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-cf7p9"] Nov 25 12:22:20 crc kubenswrapper[4715]: W1125 12:22:20.172748 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1c71c09_df63_430f_80f2_74d6cdb35e1a.slice/crio-cbaa8189969ce729fcb65109100acff5cce1ef52ea7f1211af35af585ccb61b3 WatchSource:0}: Error finding container cbaa8189969ce729fcb65109100acff5cce1ef52ea7f1211af35af585ccb61b3: Status 404 returned error can't find the container with id cbaa8189969ce729fcb65109100acff5cce1ef52ea7f1211af35af585ccb61b3 Nov 25 12:22:20 crc kubenswrapper[4715]: I1125 12:22:20.931152 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" event={"ID":"b1c71c09-df63-430f-80f2-74d6cdb35e1a","Type":"ContainerStarted","Data":"cbaa8189969ce729fcb65109100acff5cce1ef52ea7f1211af35af585ccb61b3"} Nov 25 12:22:20 crc kubenswrapper[4715]: I1125 12:22:20.933686 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"85183f2d-ee39-42f1-9153-9a7e6599ef2b","Type":"ContainerStarted","Data":"158d0088c573be7d88f24df309a6637ebf95bdef31ef90fd5cc9b610832b16b3"} Nov 25 12:22:20 crc kubenswrapper[4715]: I1125 12:22:20.936403 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"473c10ab-6523-402f-b66c-5d846a033294","Type":"ContainerStarted","Data":"a93795629aeb9a55a005482d2ad27bd9925c959dcc035e2849bd6099019e1460"} Nov 25 12:22:20 crc kubenswrapper[4715]: I1125 12:22:20.939158 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"8fb56a14-d3f7-4cf3-ae0b-8aad521a3423","Type":"ContainerStarted","Data":"acc2a0bed308489ea12ef5bee021e4e52d5880a0c3099d31dedf2567b92cc2f2"} Nov 25 12:22:20 crc kubenswrapper[4715]: I1125 12:22:20.956832 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=7.694408034 podStartE2EDuration="15.956813291s" podCreationTimestamp="2025-11-25 12:22:05 +0000 UTC" firstStartedPulling="2025-11-25 12:22:07.034365207 +0000 UTC m=+777.541868228" lastFinishedPulling="2025-11-25 12:22:15.296770464 +0000 UTC m=+785.804273485" observedRunningTime="2025-11-25 12:22:20.952740046 +0000 UTC m=+791.460243077" watchObservedRunningTime="2025-11-25 12:22:20.956813291 +0000 UTC m=+791.464316312" Nov 25 12:22:20 crc kubenswrapper[4715]: I1125 12:22:20.978010 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=7.690172967 podStartE2EDuration="15.97799106s" podCreationTimestamp="2025-11-25 12:22:05 +0000 UTC" firstStartedPulling="2025-11-25 12:22:06.962870826 +0000 UTC m=+777.470373847" lastFinishedPulling="2025-11-25 12:22:15.250688899 +0000 UTC m=+785.758191940" observedRunningTime="2025-11-25 12:22:20.97720964 +0000 UTC m=+791.484712671" watchObservedRunningTime="2025-11-25 12:22:20.97799106 +0000 UTC m=+791.485494091" Nov 25 12:22:20 crc kubenswrapper[4715]: I1125 12:22:20.997450 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=7.823438618 podStartE2EDuration="15.997433394s" podCreationTimestamp="2025-11-25 12:22:05 +0000 UTC" firstStartedPulling="2025-11-25 12:22:07.12247261 +0000 UTC m=+777.629975631" lastFinishedPulling="2025-11-25 12:22:15.296467386 +0000 UTC m=+785.803970407" observedRunningTime="2025-11-25 12:22:20.994024175 +0000 UTC m=+791.501527196" watchObservedRunningTime="2025-11-25 12:22:20.997433394 +0000 UTC m=+791.504936415" Nov 25 12:22:23 crc kubenswrapper[4715]: E1125 12:22:23.529151 4715 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.40:58160->38.129.56.40:42717: write tcp 38.129.56.40:58160->38.129.56.40:42717: write: broken pipe Nov 25 12:22:23 crc kubenswrapper[4715]: I1125 12:22:23.956975 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" event={"ID":"b1c71c09-df63-430f-80f2-74d6cdb35e1a","Type":"ContainerStarted","Data":"b428728c8bc65648fe604ab01ace82c0a7921bcdb61882596600fa1ce45dafd1"} Nov 25 12:22:23 crc kubenswrapper[4715]: I1125 12:22:23.971147 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" podStartSLOduration=1.479922224 podStartE2EDuration="4.971126609s" podCreationTimestamp="2025-11-25 12:22:19 +0000 UTC" firstStartedPulling="2025-11-25 12:22:20.175050996 +0000 UTC m=+790.682554017" lastFinishedPulling="2025-11-25 12:22:23.666255391 +0000 UTC m=+794.173758402" observedRunningTime="2025-11-25 12:22:23.968289456 +0000 UTC m=+794.475792477" watchObservedRunningTime="2025-11-25 12:22:23.971126609 +0000 UTC m=+794.478629630" Nov 25 12:22:26 crc kubenswrapper[4715]: I1125 12:22:26.600013 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:26 crc kubenswrapper[4715]: I1125 12:22:26.601321 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:26 crc kubenswrapper[4715]: I1125 12:22:26.641981 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:26 crc kubenswrapper[4715]: I1125 12:22:26.642737 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:26 crc kubenswrapper[4715]: I1125 12:22:26.642771 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:26 crc kubenswrapper[4715]: I1125 12:22:26.642783 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:26 crc kubenswrapper[4715]: I1125 12:22:26.922650 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 25 12:22:29 crc kubenswrapper[4715]: I1125 12:22:29.846312 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" Nov 25 12:22:29 crc kubenswrapper[4715]: I1125 12:22:29.846826 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" Nov 25 12:22:30 crc kubenswrapper[4715]: I1125 12:22:30.214533 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" Nov 25 12:22:30 crc kubenswrapper[4715]: I1125 12:22:30.265804 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-cf7p9" Nov 25 12:22:31 crc kubenswrapper[4715]: I1125 12:22:31.737522 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm"] Nov 25 12:22:31 crc kubenswrapper[4715]: I1125 12:22:31.738988 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:31 crc kubenswrapper[4715]: I1125 12:22:31.740490 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7xsfb" Nov 25 12:22:31 crc kubenswrapper[4715]: I1125 12:22:31.749234 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm"] Nov 25 12:22:31 crc kubenswrapper[4715]: I1125 12:22:31.921395 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:31 crc kubenswrapper[4715]: I1125 12:22:31.921481 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvbhv\" (UniqueName: \"kubernetes.io/projected/d60e6e6e-375e-4c32-be65-21d1e1d518c6-kube-api-access-zvbhv\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:31 crc kubenswrapper[4715]: I1125 12:22:31.921515 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:32 crc kubenswrapper[4715]: I1125 12:22:32.023092 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvbhv\" (UniqueName: \"kubernetes.io/projected/d60e6e6e-375e-4c32-be65-21d1e1d518c6-kube-api-access-zvbhv\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:32 crc kubenswrapper[4715]: I1125 12:22:32.023150 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:32 crc kubenswrapper[4715]: I1125 12:22:32.023254 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:32 crc kubenswrapper[4715]: I1125 12:22:32.023708 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:32 crc kubenswrapper[4715]: I1125 12:22:32.023860 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:32 crc kubenswrapper[4715]: I1125 12:22:32.046752 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvbhv\" (UniqueName: \"kubernetes.io/projected/d60e6e6e-375e-4c32-be65-21d1e1d518c6-kube-api-access-zvbhv\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:32 crc kubenswrapper[4715]: I1125 12:22:32.067343 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:32 crc kubenswrapper[4715]: I1125 12:22:32.501303 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm"] Nov 25 12:22:32 crc kubenswrapper[4715]: W1125 12:22:32.510821 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd60e6e6e_375e_4c32_be65_21d1e1d518c6.slice/crio-00c88470b8dfa0b2f172e70f29c769a121e08635d1b1a72c139aa70ec6960951 WatchSource:0}: Error finding container 00c88470b8dfa0b2f172e70f29c769a121e08635d1b1a72c139aa70ec6960951: Status 404 returned error can't find the container with id 00c88470b8dfa0b2f172e70f29c769a121e08635d1b1a72c139aa70ec6960951 Nov 25 12:22:33 crc kubenswrapper[4715]: I1125 12:22:33.256744 4715 generic.go:334] "Generic (PLEG): container finished" podID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerID="f5b994fb484ed0436470a01e905fd72dc02e4a2b607e1b5c0cdbf6b0567fcae7" exitCode=0 Nov 25 12:22:33 crc kubenswrapper[4715]: I1125 12:22:33.256790 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" event={"ID":"d60e6e6e-375e-4c32-be65-21d1e1d518c6","Type":"ContainerDied","Data":"f5b994fb484ed0436470a01e905fd72dc02e4a2b607e1b5c0cdbf6b0567fcae7"} Nov 25 12:22:33 crc kubenswrapper[4715]: I1125 12:22:33.257064 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" event={"ID":"d60e6e6e-375e-4c32-be65-21d1e1d518c6","Type":"ContainerStarted","Data":"00c88470b8dfa0b2f172e70f29c769a121e08635d1b1a72c139aa70ec6960951"} Nov 25 12:22:33 crc kubenswrapper[4715]: I1125 12:22:33.610137 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:33 crc kubenswrapper[4715]: I1125 12:22:33.673888 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 25 12:22:34 crc kubenswrapper[4715]: E1125 12:22:34.148131 4715 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.40:48772->38.129.56.40:42717: write tcp 38.129.56.40:48772->38.129.56.40:42717: write: broken pipe Nov 25 12:22:34 crc kubenswrapper[4715]: I1125 12:22:34.264863 4715 generic.go:334] "Generic (PLEG): container finished" podID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerID="c50b146f1f291c363f95e5107442f8ea2ee624fbc28a003e40f19d1d543a78cb" exitCode=0 Nov 25 12:22:34 crc kubenswrapper[4715]: I1125 12:22:34.264932 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" event={"ID":"d60e6e6e-375e-4c32-be65-21d1e1d518c6","Type":"ContainerDied","Data":"c50b146f1f291c363f95e5107442f8ea2ee624fbc28a003e40f19d1d543a78cb"} Nov 25 12:22:35 crc kubenswrapper[4715]: I1125 12:22:35.274601 4715 generic.go:334] "Generic (PLEG): container finished" podID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerID="ae3b3ec446ba2adcc35ae9688182c69ffdb8ae5e0d6e65a98d5b040e74ef6670" exitCode=0 Nov 25 12:22:35 crc kubenswrapper[4715]: I1125 12:22:35.274711 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" event={"ID":"d60e6e6e-375e-4c32-be65-21d1e1d518c6","Type":"ContainerDied","Data":"ae3b3ec446ba2adcc35ae9688182c69ffdb8ae5e0d6e65a98d5b040e74ef6670"} Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.552754 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.684370 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvbhv\" (UniqueName: \"kubernetes.io/projected/d60e6e6e-375e-4c32-be65-21d1e1d518c6-kube-api-access-zvbhv\") pod \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.684476 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-util\") pod \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.684522 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-bundle\") pod \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\" (UID: \"d60e6e6e-375e-4c32-be65-21d1e1d518c6\") " Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.685024 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-util" (OuterVolumeSpecName: "util") pod "d60e6e6e-375e-4c32-be65-21d1e1d518c6" (UID: "d60e6e6e-375e-4c32-be65-21d1e1d518c6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.685299 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-bundle" (OuterVolumeSpecName: "bundle") pod "d60e6e6e-375e-4c32-be65-21d1e1d518c6" (UID: "d60e6e6e-375e-4c32-be65-21d1e1d518c6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.690282 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d60e6e6e-375e-4c32-be65-21d1e1d518c6-kube-api-access-zvbhv" (OuterVolumeSpecName: "kube-api-access-zvbhv") pod "d60e6e6e-375e-4c32-be65-21d1e1d518c6" (UID: "d60e6e6e-375e-4c32-be65-21d1e1d518c6"). InnerVolumeSpecName "kube-api-access-zvbhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.701674 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="473c10ab-6523-402f-b66c-5d846a033294" containerName="galera" probeResult="failure" output=< Nov 25 12:22:36 crc kubenswrapper[4715]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Nov 25 12:22:36 crc kubenswrapper[4715]: > Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.786194 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvbhv\" (UniqueName: \"kubernetes.io/projected/d60e6e6e-375e-4c32-be65-21d1e1d518c6-kube-api-access-zvbhv\") on node \"crc\" DevicePath \"\"" Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.786225 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-util\") on node \"crc\" DevicePath \"\"" Nov 25 12:22:36 crc kubenswrapper[4715]: I1125 12:22:36.786235 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d60e6e6e-375e-4c32-be65-21d1e1d518c6-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:22:37 crc kubenswrapper[4715]: I1125 12:22:37.295495 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" event={"ID":"d60e6e6e-375e-4c32-be65-21d1e1d518c6","Type":"ContainerDied","Data":"00c88470b8dfa0b2f172e70f29c769a121e08635d1b1a72c139aa70ec6960951"} Nov 25 12:22:37 crc kubenswrapper[4715]: I1125 12:22:37.295947 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00c88470b8dfa0b2f172e70f29c769a121e08635d1b1a72c139aa70ec6960951" Nov 25 12:22:37 crc kubenswrapper[4715]: I1125 12:22:37.295549 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm" Nov 25 12:22:42 crc kubenswrapper[4715]: I1125 12:22:42.444688 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:42 crc kubenswrapper[4715]: I1125 12:22:42.516547 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 25 12:22:47 crc kubenswrapper[4715]: I1125 12:22:47.250341 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:47 crc kubenswrapper[4715]: I1125 12:22:47.315522 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 25 12:22:47 crc kubenswrapper[4715]: I1125 12:22:47.943636 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:22:47 crc kubenswrapper[4715]: I1125 12:22:47.943734 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.675794 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j"] Nov 25 12:22:48 crc kubenswrapper[4715]: E1125 12:22:48.676082 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerName="extract" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.676097 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerName="extract" Nov 25 12:22:48 crc kubenswrapper[4715]: E1125 12:22:48.676115 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerName="util" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.676122 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerName="util" Nov 25 12:22:48 crc kubenswrapper[4715]: E1125 12:22:48.676139 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerName="pull" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.676148 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerName="pull" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.676295 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d60e6e6e-375e-4c32-be65-21d1e1d518c6" containerName="extract" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.676763 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.679570 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-hbz9w" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.685375 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j"] Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.757323 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64tqf\" (UniqueName: \"kubernetes.io/projected/0508bead-e249-4ac2-84d1-741dcf4030c7-kube-api-access-64tqf\") pod \"rabbitmq-cluster-operator-779fc9694b-qsn2j\" (UID: \"0508bead-e249-4ac2-84d1-741dcf4030c7\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.858776 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64tqf\" (UniqueName: \"kubernetes.io/projected/0508bead-e249-4ac2-84d1-741dcf4030c7-kube-api-access-64tqf\") pod \"rabbitmq-cluster-operator-779fc9694b-qsn2j\" (UID: \"0508bead-e249-4ac2-84d1-741dcf4030c7\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.893304 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64tqf\" (UniqueName: \"kubernetes.io/projected/0508bead-e249-4ac2-84d1-741dcf4030c7-kube-api-access-64tqf\") pod \"rabbitmq-cluster-operator-779fc9694b-qsn2j\" (UID: \"0508bead-e249-4ac2-84d1-741dcf4030c7\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j" Nov 25 12:22:48 crc kubenswrapper[4715]: I1125 12:22:48.993055 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j" Nov 25 12:22:49 crc kubenswrapper[4715]: I1125 12:22:49.426558 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j"] Nov 25 12:22:50 crc kubenswrapper[4715]: I1125 12:22:50.368666 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j" event={"ID":"0508bead-e249-4ac2-84d1-741dcf4030c7","Type":"ContainerStarted","Data":"5bf16a597ef72f3f92f6b84d49d53d374f16512ce561379bfc74bc2c677e4a99"} Nov 25 12:22:52 crc kubenswrapper[4715]: I1125 12:22:52.381769 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j" event={"ID":"0508bead-e249-4ac2-84d1-741dcf4030c7","Type":"ContainerStarted","Data":"2a0d847cda0d40446398950732025c69e0ad5280de8ca25566e4216cde3390c8"} Nov 25 12:22:52 crc kubenswrapper[4715]: I1125 12:22:52.400312 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-qsn2j" podStartSLOduration=1.643567955 podStartE2EDuration="4.40028295s" podCreationTimestamp="2025-11-25 12:22:48 +0000 UTC" firstStartedPulling="2025-11-25 12:22:49.43457805 +0000 UTC m=+819.942081071" lastFinishedPulling="2025-11-25 12:22:52.191293045 +0000 UTC m=+822.698796066" observedRunningTime="2025-11-25 12:22:52.398470673 +0000 UTC m=+822.905973724" watchObservedRunningTime="2025-11-25 12:22:52.40028295 +0000 UTC m=+822.907786031" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.524227 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.525890 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.528568 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.528893 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.528956 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.529406 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.529892 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-ql469" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.543139 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.589097 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42f62681-9aba-4c32-af6b-0ec4e1503b65-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.589156 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42f62681-9aba-4c32-af6b-0ec4e1503b65-pod-info\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.589276 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3b635c3d-b49d-4047-9296-a7dadc87ead3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b635c3d-b49d-4047-9296-a7dadc87ead3\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.589316 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42f62681-9aba-4c32-af6b-0ec4e1503b65-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.589330 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42f62681-9aba-4c32-af6b-0ec4e1503b65-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.589345 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42f62681-9aba-4c32-af6b-0ec4e1503b65-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.589364 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42f62681-9aba-4c32-af6b-0ec4e1503b65-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.589384 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pjhv\" (UniqueName: \"kubernetes.io/projected/42f62681-9aba-4c32-af6b-0ec4e1503b65-kube-api-access-6pjhv\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.690964 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3b635c3d-b49d-4047-9296-a7dadc87ead3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b635c3d-b49d-4047-9296-a7dadc87ead3\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.691137 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42f62681-9aba-4c32-af6b-0ec4e1503b65-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.691167 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42f62681-9aba-4c32-af6b-0ec4e1503b65-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.691207 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42f62681-9aba-4c32-af6b-0ec4e1503b65-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.691234 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42f62681-9aba-4c32-af6b-0ec4e1503b65-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.691262 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pjhv\" (UniqueName: \"kubernetes.io/projected/42f62681-9aba-4c32-af6b-0ec4e1503b65-kube-api-access-6pjhv\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.691297 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42f62681-9aba-4c32-af6b-0ec4e1503b65-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.691332 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42f62681-9aba-4c32-af6b-0ec4e1503b65-pod-info\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.692001 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/42f62681-9aba-4c32-af6b-0ec4e1503b65-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.692259 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/42f62681-9aba-4c32-af6b-0ec4e1503b65-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.692897 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/42f62681-9aba-4c32-af6b-0ec4e1503b65-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.697326 4715 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.697591 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3b635c3d-b49d-4047-9296-a7dadc87ead3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b635c3d-b49d-4047-9296-a7dadc87ead3\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0a3768f9332c10b36f3bf31abcccc0f051a8f6ffcb973eff5f59d444cb87fe73/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.704408 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/42f62681-9aba-4c32-af6b-0ec4e1503b65-pod-info\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.705268 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/42f62681-9aba-4c32-af6b-0ec4e1503b65-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.705744 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/42f62681-9aba-4c32-af6b-0ec4e1503b65-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.711627 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pjhv\" (UniqueName: \"kubernetes.io/projected/42f62681-9aba-4c32-af6b-0ec4e1503b65-kube-api-access-6pjhv\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.727044 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3b635c3d-b49d-4047-9296-a7dadc87ead3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b635c3d-b49d-4047-9296-a7dadc87ead3\") pod \"rabbitmq-server-0\" (UID: \"42f62681-9aba-4c32-af6b-0ec4e1503b65\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:58 crc kubenswrapper[4715]: I1125 12:22:58.855521 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:22:59 crc kubenswrapper[4715]: I1125 12:22:59.327741 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 25 12:22:59 crc kubenswrapper[4715]: I1125 12:22:59.428090 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"42f62681-9aba-4c32-af6b-0ec4e1503b65","Type":"ContainerStarted","Data":"eb1262a84623daf243f1cd4f7993d6f3f50b5c82436e4371d35196e34c4e02a1"} Nov 25 12:23:00 crc kubenswrapper[4715]: I1125 12:23:00.104698 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-2dlh4"] Nov 25 12:23:00 crc kubenswrapper[4715]: I1125 12:23:00.105939 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-2dlh4" Nov 25 12:23:00 crc kubenswrapper[4715]: I1125 12:23:00.108440 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-xxhlf" Nov 25 12:23:00 crc kubenswrapper[4715]: I1125 12:23:00.112208 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-2dlh4"] Nov 25 12:23:00 crc kubenswrapper[4715]: I1125 12:23:00.213476 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p796m\" (UniqueName: \"kubernetes.io/projected/805596fb-95a8-4bfe-b3ba-64abfe06a16b-kube-api-access-p796m\") pod \"keystone-operator-index-2dlh4\" (UID: \"805596fb-95a8-4bfe-b3ba-64abfe06a16b\") " pod="openstack-operators/keystone-operator-index-2dlh4" Nov 25 12:23:00 crc kubenswrapper[4715]: I1125 12:23:00.314733 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p796m\" (UniqueName: \"kubernetes.io/projected/805596fb-95a8-4bfe-b3ba-64abfe06a16b-kube-api-access-p796m\") pod \"keystone-operator-index-2dlh4\" (UID: \"805596fb-95a8-4bfe-b3ba-64abfe06a16b\") " pod="openstack-operators/keystone-operator-index-2dlh4" Nov 25 12:23:00 crc kubenswrapper[4715]: I1125 12:23:00.335572 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p796m\" (UniqueName: \"kubernetes.io/projected/805596fb-95a8-4bfe-b3ba-64abfe06a16b-kube-api-access-p796m\") pod \"keystone-operator-index-2dlh4\" (UID: \"805596fb-95a8-4bfe-b3ba-64abfe06a16b\") " pod="openstack-operators/keystone-operator-index-2dlh4" Nov 25 12:23:00 crc kubenswrapper[4715]: I1125 12:23:00.438312 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-2dlh4" Nov 25 12:23:00 crc kubenswrapper[4715]: I1125 12:23:00.889073 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-2dlh4"] Nov 25 12:23:01 crc kubenswrapper[4715]: I1125 12:23:01.452860 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-2dlh4" event={"ID":"805596fb-95a8-4bfe-b3ba-64abfe06a16b","Type":"ContainerStarted","Data":"97b80b2d7cc457d40d89550ee655f72189a0c8bc593e689305953decc1a0bbaf"} Nov 25 12:23:02 crc kubenswrapper[4715]: I1125 12:23:02.459753 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-2dlh4" event={"ID":"805596fb-95a8-4bfe-b3ba-64abfe06a16b","Type":"ContainerStarted","Data":"8f1c0484cf7899b8437ad642fadd22054ca4d65f1dfd3419ca725777d6a39d29"} Nov 25 12:23:02 crc kubenswrapper[4715]: I1125 12:23:02.476646 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-2dlh4" podStartSLOduration=1.73747317 podStartE2EDuration="2.476631851s" podCreationTimestamp="2025-11-25 12:23:00 +0000 UTC" firstStartedPulling="2025-11-25 12:23:00.908093371 +0000 UTC m=+831.415596412" lastFinishedPulling="2025-11-25 12:23:01.647252032 +0000 UTC m=+832.154755093" observedRunningTime="2025-11-25 12:23:02.473591121 +0000 UTC m=+832.981094152" watchObservedRunningTime="2025-11-25 12:23:02.476631851 +0000 UTC m=+832.984134872" Nov 25 12:23:04 crc kubenswrapper[4715]: I1125 12:23:04.301796 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-2dlh4"] Nov 25 12:23:04 crc kubenswrapper[4715]: I1125 12:23:04.471603 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-2dlh4" podUID="805596fb-95a8-4bfe-b3ba-64abfe06a16b" containerName="registry-server" containerID="cri-o://8f1c0484cf7899b8437ad642fadd22054ca4d65f1dfd3419ca725777d6a39d29" gracePeriod=2 Nov 25 12:23:04 crc kubenswrapper[4715]: I1125 12:23:04.907994 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-4sskb"] Nov 25 12:23:04 crc kubenswrapper[4715]: I1125 12:23:04.908707 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-4sskb" Nov 25 12:23:04 crc kubenswrapper[4715]: I1125 12:23:04.920283 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-4sskb"] Nov 25 12:23:04 crc kubenswrapper[4715]: I1125 12:23:04.981296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66jhf\" (UniqueName: \"kubernetes.io/projected/c69a3c36-24a6-4871-8d82-a811e2ed5afc-kube-api-access-66jhf\") pod \"keystone-operator-index-4sskb\" (UID: \"c69a3c36-24a6-4871-8d82-a811e2ed5afc\") " pod="openstack-operators/keystone-operator-index-4sskb" Nov 25 12:23:05 crc kubenswrapper[4715]: I1125 12:23:05.082642 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66jhf\" (UniqueName: \"kubernetes.io/projected/c69a3c36-24a6-4871-8d82-a811e2ed5afc-kube-api-access-66jhf\") pod \"keystone-operator-index-4sskb\" (UID: \"c69a3c36-24a6-4871-8d82-a811e2ed5afc\") " pod="openstack-operators/keystone-operator-index-4sskb" Nov 25 12:23:05 crc kubenswrapper[4715]: I1125 12:23:05.108764 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66jhf\" (UniqueName: \"kubernetes.io/projected/c69a3c36-24a6-4871-8d82-a811e2ed5afc-kube-api-access-66jhf\") pod \"keystone-operator-index-4sskb\" (UID: \"c69a3c36-24a6-4871-8d82-a811e2ed5afc\") " pod="openstack-operators/keystone-operator-index-4sskb" Nov 25 12:23:05 crc kubenswrapper[4715]: I1125 12:23:05.234213 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-4sskb" Nov 25 12:23:05 crc kubenswrapper[4715]: I1125 12:23:05.478469 4715 generic.go:334] "Generic (PLEG): container finished" podID="805596fb-95a8-4bfe-b3ba-64abfe06a16b" containerID="8f1c0484cf7899b8437ad642fadd22054ca4d65f1dfd3419ca725777d6a39d29" exitCode=0 Nov 25 12:23:05 crc kubenswrapper[4715]: I1125 12:23:05.478519 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-2dlh4" event={"ID":"805596fb-95a8-4bfe-b3ba-64abfe06a16b","Type":"ContainerDied","Data":"8f1c0484cf7899b8437ad642fadd22054ca4d65f1dfd3419ca725777d6a39d29"} Nov 25 12:23:05 crc kubenswrapper[4715]: I1125 12:23:05.663366 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-2dlh4" Nov 25 12:23:05 crc kubenswrapper[4715]: I1125 12:23:05.792545 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p796m\" (UniqueName: \"kubernetes.io/projected/805596fb-95a8-4bfe-b3ba-64abfe06a16b-kube-api-access-p796m\") pod \"805596fb-95a8-4bfe-b3ba-64abfe06a16b\" (UID: \"805596fb-95a8-4bfe-b3ba-64abfe06a16b\") " Nov 25 12:23:05 crc kubenswrapper[4715]: I1125 12:23:05.796726 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/805596fb-95a8-4bfe-b3ba-64abfe06a16b-kube-api-access-p796m" (OuterVolumeSpecName: "kube-api-access-p796m") pod "805596fb-95a8-4bfe-b3ba-64abfe06a16b" (UID: "805596fb-95a8-4bfe-b3ba-64abfe06a16b"). InnerVolumeSpecName "kube-api-access-p796m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:23:05 crc kubenswrapper[4715]: I1125 12:23:05.895247 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p796m\" (UniqueName: \"kubernetes.io/projected/805596fb-95a8-4bfe-b3ba-64abfe06a16b-kube-api-access-p796m\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:06 crc kubenswrapper[4715]: I1125 12:23:06.018340 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-4sskb"] Nov 25 12:23:06 crc kubenswrapper[4715]: W1125 12:23:06.021900 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc69a3c36_24a6_4871_8d82_a811e2ed5afc.slice/crio-d697f3a7463a3f4ee8b4e446a58f352110f8d579527c63b524a5ecce36ceaaf4 WatchSource:0}: Error finding container d697f3a7463a3f4ee8b4e446a58f352110f8d579527c63b524a5ecce36ceaaf4: Status 404 returned error can't find the container with id d697f3a7463a3f4ee8b4e446a58f352110f8d579527c63b524a5ecce36ceaaf4 Nov 25 12:23:06 crc kubenswrapper[4715]: I1125 12:23:06.486816 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-4sskb" event={"ID":"c69a3c36-24a6-4871-8d82-a811e2ed5afc","Type":"ContainerStarted","Data":"d697f3a7463a3f4ee8b4e446a58f352110f8d579527c63b524a5ecce36ceaaf4"} Nov 25 12:23:06 crc kubenswrapper[4715]: I1125 12:23:06.489676 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-2dlh4" Nov 25 12:23:06 crc kubenswrapper[4715]: I1125 12:23:06.489576 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-2dlh4" event={"ID":"805596fb-95a8-4bfe-b3ba-64abfe06a16b","Type":"ContainerDied","Data":"97b80b2d7cc457d40d89550ee655f72189a0c8bc593e689305953decc1a0bbaf"} Nov 25 12:23:06 crc kubenswrapper[4715]: I1125 12:23:06.490096 4715 scope.go:117] "RemoveContainer" containerID="8f1c0484cf7899b8437ad642fadd22054ca4d65f1dfd3419ca725777d6a39d29" Nov 25 12:23:06 crc kubenswrapper[4715]: I1125 12:23:06.542598 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-2dlh4"] Nov 25 12:23:06 crc kubenswrapper[4715]: I1125 12:23:06.547684 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-2dlh4"] Nov 25 12:23:06 crc kubenswrapper[4715]: I1125 12:23:06.702904 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="805596fb-95a8-4bfe-b3ba-64abfe06a16b" path="/var/lib/kubelet/pods/805596fb-95a8-4bfe-b3ba-64abfe06a16b/volumes" Nov 25 12:23:07 crc kubenswrapper[4715]: I1125 12:23:07.504372 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-4sskb" event={"ID":"c69a3c36-24a6-4871-8d82-a811e2ed5afc","Type":"ContainerStarted","Data":"47f9f83b5fb9dff970221980509a32cef39661da6e0f2c44f0a1f27139ab4f2a"} Nov 25 12:23:07 crc kubenswrapper[4715]: I1125 12:23:07.506130 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"42f62681-9aba-4c32-af6b-0ec4e1503b65","Type":"ContainerStarted","Data":"936f8174340c0a894e36d03629d38cba58fa215492b8679c810ac200e465a12a"} Nov 25 12:23:07 crc kubenswrapper[4715]: I1125 12:23:07.527177 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-4sskb" podStartSLOduration=3.116839294 podStartE2EDuration="3.527145785s" podCreationTimestamp="2025-11-25 12:23:04 +0000 UTC" firstStartedPulling="2025-11-25 12:23:06.026246998 +0000 UTC m=+836.533750019" lastFinishedPulling="2025-11-25 12:23:06.436553449 +0000 UTC m=+836.944056510" observedRunningTime="2025-11-25 12:23:07.524345452 +0000 UTC m=+838.031848543" watchObservedRunningTime="2025-11-25 12:23:07.527145785 +0000 UTC m=+838.034648856" Nov 25 12:23:15 crc kubenswrapper[4715]: I1125 12:23:15.234935 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-4sskb" Nov 25 12:23:15 crc kubenswrapper[4715]: I1125 12:23:15.235085 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-4sskb" Nov 25 12:23:15 crc kubenswrapper[4715]: I1125 12:23:15.265430 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-4sskb" Nov 25 12:23:15 crc kubenswrapper[4715]: I1125 12:23:15.593041 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-4sskb" Nov 25 12:23:17 crc kubenswrapper[4715]: I1125 12:23:17.944311 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:23:17 crc kubenswrapper[4715]: I1125 12:23:17.944863 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:23:17 crc kubenswrapper[4715]: I1125 12:23:17.944921 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:23:17 crc kubenswrapper[4715]: I1125 12:23:17.945619 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"23a04d777596661d2024d01da087796f35f637e5c71480856ee076f6f5b38b7a"} pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:23:17 crc kubenswrapper[4715]: I1125 12:23:17.945710 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" containerID="cri-o://23a04d777596661d2024d01da087796f35f637e5c71480856ee076f6f5b38b7a" gracePeriod=600 Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.157826 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95"] Nov 25 12:23:18 crc kubenswrapper[4715]: E1125 12:23:18.158111 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="805596fb-95a8-4bfe-b3ba-64abfe06a16b" containerName="registry-server" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.158122 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="805596fb-95a8-4bfe-b3ba-64abfe06a16b" containerName="registry-server" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.158273 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="805596fb-95a8-4bfe-b3ba-64abfe06a16b" containerName="registry-server" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.159250 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.161983 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7xsfb" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.172785 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95"] Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.266507 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-util\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.266558 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krpg9\" (UniqueName: \"kubernetes.io/projected/27922bfd-ecc0-49b9-b1f5-72195ceac78d-kube-api-access-krpg9\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.266582 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-bundle\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.368353 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-util\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.368879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krpg9\" (UniqueName: \"kubernetes.io/projected/27922bfd-ecc0-49b9-b1f5-72195ceac78d-kube-api-access-krpg9\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.368942 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-bundle\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.370079 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-bundle\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.370293 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-util\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.398816 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krpg9\" (UniqueName: \"kubernetes.io/projected/27922bfd-ecc0-49b9-b1f5-72195ceac78d-kube-api-access-krpg9\") pod \"62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.478552 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.575793 4715 generic.go:334] "Generic (PLEG): container finished" podID="51676837-adc4-4424-b527-920a6528b6a2" containerID="23a04d777596661d2024d01da087796f35f637e5c71480856ee076f6f5b38b7a" exitCode=0 Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.575832 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerDied","Data":"23a04d777596661d2024d01da087796f35f637e5c71480856ee076f6f5b38b7a"} Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.575862 4715 scope.go:117] "RemoveContainer" containerID="d16257d5ac23a2866262558b39c2b844ea7649b0c06fa81698cc8fe1fd3c8f95" Nov 25 12:23:18 crc kubenswrapper[4715]: I1125 12:23:18.672873 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95"] Nov 25 12:23:18 crc kubenswrapper[4715]: W1125 12:23:18.677403 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27922bfd_ecc0_49b9_b1f5_72195ceac78d.slice/crio-af3cd3f1bc57a6586c4b2fffaf7d78172cc8bdbce9ff141298ca6a03981c6b1d WatchSource:0}: Error finding container af3cd3f1bc57a6586c4b2fffaf7d78172cc8bdbce9ff141298ca6a03981c6b1d: Status 404 returned error can't find the container with id af3cd3f1bc57a6586c4b2fffaf7d78172cc8bdbce9ff141298ca6a03981c6b1d Nov 25 12:23:19 crc kubenswrapper[4715]: I1125 12:23:19.584166 4715 generic.go:334] "Generic (PLEG): container finished" podID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerID="24bf6508b1c4b91b5e9ab6bbbc688ebd5f1ca57caf7e2abb191f662fc5f97022" exitCode=0 Nov 25 12:23:19 crc kubenswrapper[4715]: I1125 12:23:19.584263 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" event={"ID":"27922bfd-ecc0-49b9-b1f5-72195ceac78d","Type":"ContainerDied","Data":"24bf6508b1c4b91b5e9ab6bbbc688ebd5f1ca57caf7e2abb191f662fc5f97022"} Nov 25 12:23:19 crc kubenswrapper[4715]: I1125 12:23:19.584613 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" event={"ID":"27922bfd-ecc0-49b9-b1f5-72195ceac78d","Type":"ContainerStarted","Data":"af3cd3f1bc57a6586c4b2fffaf7d78172cc8bdbce9ff141298ca6a03981c6b1d"} Nov 25 12:23:19 crc kubenswrapper[4715]: I1125 12:23:19.589159 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"244d5332aa55fb80cd40f8e16728cd8f4bf2022a5bf11e13456cf894fe4243d1"} Nov 25 12:23:20 crc kubenswrapper[4715]: I1125 12:23:20.598451 4715 generic.go:334] "Generic (PLEG): container finished" podID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerID="eee1a188556dc78ec211d8151c9e645bdd16b19f33018d7f76d26eb8bb91171e" exitCode=0 Nov 25 12:23:20 crc kubenswrapper[4715]: I1125 12:23:20.598553 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" event={"ID":"27922bfd-ecc0-49b9-b1f5-72195ceac78d","Type":"ContainerDied","Data":"eee1a188556dc78ec211d8151c9e645bdd16b19f33018d7f76d26eb8bb91171e"} Nov 25 12:23:21 crc kubenswrapper[4715]: I1125 12:23:21.606818 4715 generic.go:334] "Generic (PLEG): container finished" podID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerID="4a6425d46e2fbe339ab57b09c631ac4f833cba16aac15a167ad9d63182e2362b" exitCode=0 Nov 25 12:23:21 crc kubenswrapper[4715]: I1125 12:23:21.606877 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" event={"ID":"27922bfd-ecc0-49b9-b1f5-72195ceac78d","Type":"ContainerDied","Data":"4a6425d46e2fbe339ab57b09c631ac4f833cba16aac15a167ad9d63182e2362b"} Nov 25 12:23:22 crc kubenswrapper[4715]: I1125 12:23:22.865860 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:22 crc kubenswrapper[4715]: I1125 12:23:22.930692 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-bundle\") pod \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " Nov 25 12:23:22 crc kubenswrapper[4715]: I1125 12:23:22.930834 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-util\") pod \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " Nov 25 12:23:22 crc kubenswrapper[4715]: I1125 12:23:22.930883 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krpg9\" (UniqueName: \"kubernetes.io/projected/27922bfd-ecc0-49b9-b1f5-72195ceac78d-kube-api-access-krpg9\") pod \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\" (UID: \"27922bfd-ecc0-49b9-b1f5-72195ceac78d\") " Nov 25 12:23:22 crc kubenswrapper[4715]: I1125 12:23:22.932403 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-bundle" (OuterVolumeSpecName: "bundle") pod "27922bfd-ecc0-49b9-b1f5-72195ceac78d" (UID: "27922bfd-ecc0-49b9-b1f5-72195ceac78d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:23:22 crc kubenswrapper[4715]: I1125 12:23:22.938632 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27922bfd-ecc0-49b9-b1f5-72195ceac78d-kube-api-access-krpg9" (OuterVolumeSpecName: "kube-api-access-krpg9") pod "27922bfd-ecc0-49b9-b1f5-72195ceac78d" (UID: "27922bfd-ecc0-49b9-b1f5-72195ceac78d"). InnerVolumeSpecName "kube-api-access-krpg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:23:22 crc kubenswrapper[4715]: I1125 12:23:22.947661 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-util" (OuterVolumeSpecName: "util") pod "27922bfd-ecc0-49b9-b1f5-72195ceac78d" (UID: "27922bfd-ecc0-49b9-b1f5-72195ceac78d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:23:23 crc kubenswrapper[4715]: I1125 12:23:23.032916 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-util\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:23 crc kubenswrapper[4715]: I1125 12:23:23.032960 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krpg9\" (UniqueName: \"kubernetes.io/projected/27922bfd-ecc0-49b9-b1f5-72195ceac78d-kube-api-access-krpg9\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:23 crc kubenswrapper[4715]: I1125 12:23:23.032975 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/27922bfd-ecc0-49b9-b1f5-72195ceac78d-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:23 crc kubenswrapper[4715]: I1125 12:23:23.621146 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" event={"ID":"27922bfd-ecc0-49b9-b1f5-72195ceac78d","Type":"ContainerDied","Data":"af3cd3f1bc57a6586c4b2fffaf7d78172cc8bdbce9ff141298ca6a03981c6b1d"} Nov 25 12:23:23 crc kubenswrapper[4715]: I1125 12:23:23.621235 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af3cd3f1bc57a6586c4b2fffaf7d78172cc8bdbce9ff141298ca6a03981c6b1d" Nov 25 12:23:23 crc kubenswrapper[4715]: I1125 12:23:23.621268 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.867931 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr"] Nov 25 12:23:30 crc kubenswrapper[4715]: E1125 12:23:30.868774 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerName="pull" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.868787 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerName="pull" Nov 25 12:23:30 crc kubenswrapper[4715]: E1125 12:23:30.868801 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerName="util" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.868808 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerName="util" Nov 25 12:23:30 crc kubenswrapper[4715]: E1125 12:23:30.868823 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerName="extract" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.868832 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerName="extract" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.868970 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="27922bfd-ecc0-49b9-b1f5-72195ceac78d" containerName="extract" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.869473 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.874124 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.875306 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-nldph" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.883137 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr"] Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.939435 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92dc9\" (UniqueName: \"kubernetes.io/projected/9a3e4772-2d60-4498-9cdb-0a6413b4b200-kube-api-access-92dc9\") pod \"keystone-operator-controller-manager-5b9d7554bc-4rbfr\" (UID: \"9a3e4772-2d60-4498-9cdb-0a6413b4b200\") " pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.939489 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a3e4772-2d60-4498-9cdb-0a6413b4b200-webhook-cert\") pod \"keystone-operator-controller-manager-5b9d7554bc-4rbfr\" (UID: \"9a3e4772-2d60-4498-9cdb-0a6413b4b200\") " pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:30 crc kubenswrapper[4715]: I1125 12:23:30.939528 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a3e4772-2d60-4498-9cdb-0a6413b4b200-apiservice-cert\") pod \"keystone-operator-controller-manager-5b9d7554bc-4rbfr\" (UID: \"9a3e4772-2d60-4498-9cdb-0a6413b4b200\") " pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:31 crc kubenswrapper[4715]: I1125 12:23:31.041203 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92dc9\" (UniqueName: \"kubernetes.io/projected/9a3e4772-2d60-4498-9cdb-0a6413b4b200-kube-api-access-92dc9\") pod \"keystone-operator-controller-manager-5b9d7554bc-4rbfr\" (UID: \"9a3e4772-2d60-4498-9cdb-0a6413b4b200\") " pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:31 crc kubenswrapper[4715]: I1125 12:23:31.041275 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a3e4772-2d60-4498-9cdb-0a6413b4b200-webhook-cert\") pod \"keystone-operator-controller-manager-5b9d7554bc-4rbfr\" (UID: \"9a3e4772-2d60-4498-9cdb-0a6413b4b200\") " pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:31 crc kubenswrapper[4715]: I1125 12:23:31.041324 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a3e4772-2d60-4498-9cdb-0a6413b4b200-apiservice-cert\") pod \"keystone-operator-controller-manager-5b9d7554bc-4rbfr\" (UID: \"9a3e4772-2d60-4498-9cdb-0a6413b4b200\") " pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:31 crc kubenswrapper[4715]: I1125 12:23:31.047915 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9a3e4772-2d60-4498-9cdb-0a6413b4b200-webhook-cert\") pod \"keystone-operator-controller-manager-5b9d7554bc-4rbfr\" (UID: \"9a3e4772-2d60-4498-9cdb-0a6413b4b200\") " pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:31 crc kubenswrapper[4715]: I1125 12:23:31.048039 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9a3e4772-2d60-4498-9cdb-0a6413b4b200-apiservice-cert\") pod \"keystone-operator-controller-manager-5b9d7554bc-4rbfr\" (UID: \"9a3e4772-2d60-4498-9cdb-0a6413b4b200\") " pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:31 crc kubenswrapper[4715]: I1125 12:23:31.063768 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92dc9\" (UniqueName: \"kubernetes.io/projected/9a3e4772-2d60-4498-9cdb-0a6413b4b200-kube-api-access-92dc9\") pod \"keystone-operator-controller-manager-5b9d7554bc-4rbfr\" (UID: \"9a3e4772-2d60-4498-9cdb-0a6413b4b200\") " pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:31 crc kubenswrapper[4715]: I1125 12:23:31.189202 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:31 crc kubenswrapper[4715]: I1125 12:23:31.599049 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr"] Nov 25 12:23:31 crc kubenswrapper[4715]: I1125 12:23:31.666044 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" event={"ID":"9a3e4772-2d60-4498-9cdb-0a6413b4b200","Type":"ContainerStarted","Data":"74d874dbe88a63f2d49e29e3373877f0902132468438e6d32f4c703d823e65b6"} Nov 25 12:23:36 crc kubenswrapper[4715]: I1125 12:23:36.706818 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" event={"ID":"9a3e4772-2d60-4498-9cdb-0a6413b4b200","Type":"ContainerStarted","Data":"b00205b2175bae8812b01a13362f34280a2f8c8a0ca6b73a26597d458d641226"} Nov 25 12:23:36 crc kubenswrapper[4715]: I1125 12:23:36.707353 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:36 crc kubenswrapper[4715]: I1125 12:23:36.731598 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" podStartSLOduration=2.7729638960000003 podStartE2EDuration="6.731583777s" podCreationTimestamp="2025-11-25 12:23:30 +0000 UTC" firstStartedPulling="2025-11-25 12:23:31.604888782 +0000 UTC m=+862.112391803" lastFinishedPulling="2025-11-25 12:23:35.563508663 +0000 UTC m=+866.071011684" observedRunningTime="2025-11-25 12:23:36.730199919 +0000 UTC m=+867.237702960" watchObservedRunningTime="2025-11-25 12:23:36.731583777 +0000 UTC m=+867.239086798" Nov 25 12:23:38 crc kubenswrapper[4715]: I1125 12:23:38.719608 4715 generic.go:334] "Generic (PLEG): container finished" podID="42f62681-9aba-4c32-af6b-0ec4e1503b65" containerID="936f8174340c0a894e36d03629d38cba58fa215492b8679c810ac200e465a12a" exitCode=0 Nov 25 12:23:38 crc kubenswrapper[4715]: I1125 12:23:38.719731 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"42f62681-9aba-4c32-af6b-0ec4e1503b65","Type":"ContainerDied","Data":"936f8174340c0a894e36d03629d38cba58fa215492b8679c810ac200e465a12a"} Nov 25 12:23:39 crc kubenswrapper[4715]: I1125 12:23:39.729390 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"42f62681-9aba-4c32-af6b-0ec4e1503b65","Type":"ContainerStarted","Data":"3810cf31fb130c917b1e3d33a9f47385c142030354ae2a22e45b4e91698680a5"} Nov 25 12:23:39 crc kubenswrapper[4715]: I1125 12:23:39.730089 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:23:39 crc kubenswrapper[4715]: I1125 12:23:39.754488 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.417772496 podStartE2EDuration="42.754471845s" podCreationTimestamp="2025-11-25 12:22:57 +0000 UTC" firstStartedPulling="2025-11-25 12:22:59.338110594 +0000 UTC m=+829.845613655" lastFinishedPulling="2025-11-25 12:23:05.674809963 +0000 UTC m=+836.182313004" observedRunningTime="2025-11-25 12:23:39.754090635 +0000 UTC m=+870.261593656" watchObservedRunningTime="2025-11-25 12:23:39.754471845 +0000 UTC m=+870.261974866" Nov 25 12:23:41 crc kubenswrapper[4715]: I1125 12:23:41.199606 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5b9d7554bc-4rbfr" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.240248 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-4rz8m"] Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.241244 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.248662 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl"] Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.250946 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.252986 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.255863 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl"] Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.260337 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-4rz8m"] Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.311174 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2d64e67-9d9c-46c1-855e-05f6343e223e-operator-scripts\") pod \"keystone-db-create-4rz8m\" (UID: \"d2d64e67-9d9c-46c1-855e-05f6343e223e\") " pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.311468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z5jm\" (UniqueName: \"kubernetes.io/projected/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-kube-api-access-2z5jm\") pod \"keystone-dc20-account-create-update-r7gcl\" (UID: \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\") " pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.311567 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-operator-scripts\") pod \"keystone-dc20-account-create-update-r7gcl\" (UID: \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\") " pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.311660 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j4pc\" (UniqueName: \"kubernetes.io/projected/d2d64e67-9d9c-46c1-855e-05f6343e223e-kube-api-access-5j4pc\") pod \"keystone-db-create-4rz8m\" (UID: \"d2d64e67-9d9c-46c1-855e-05f6343e223e\") " pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.413418 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2d64e67-9d9c-46c1-855e-05f6343e223e-operator-scripts\") pod \"keystone-db-create-4rz8m\" (UID: \"d2d64e67-9d9c-46c1-855e-05f6343e223e\") " pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.413732 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z5jm\" (UniqueName: \"kubernetes.io/projected/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-kube-api-access-2z5jm\") pod \"keystone-dc20-account-create-update-r7gcl\" (UID: \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\") " pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.413990 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-operator-scripts\") pod \"keystone-dc20-account-create-update-r7gcl\" (UID: \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\") " pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.414091 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j4pc\" (UniqueName: \"kubernetes.io/projected/d2d64e67-9d9c-46c1-855e-05f6343e223e-kube-api-access-5j4pc\") pod \"keystone-db-create-4rz8m\" (UID: \"d2d64e67-9d9c-46c1-855e-05f6343e223e\") " pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.415040 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2d64e67-9d9c-46c1-855e-05f6343e223e-operator-scripts\") pod \"keystone-db-create-4rz8m\" (UID: \"d2d64e67-9d9c-46c1-855e-05f6343e223e\") " pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.415040 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-operator-scripts\") pod \"keystone-dc20-account-create-update-r7gcl\" (UID: \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\") " pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.440170 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j4pc\" (UniqueName: \"kubernetes.io/projected/d2d64e67-9d9c-46c1-855e-05f6343e223e-kube-api-access-5j4pc\") pod \"keystone-db-create-4rz8m\" (UID: \"d2d64e67-9d9c-46c1-855e-05f6343e223e\") " pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.440753 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z5jm\" (UniqueName: \"kubernetes.io/projected/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-kube-api-access-2z5jm\") pod \"keystone-dc20-account-create-update-r7gcl\" (UID: \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\") " pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.566219 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:42 crc kubenswrapper[4715]: I1125 12:23:42.582385 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:43 crc kubenswrapper[4715]: I1125 12:23:43.062367 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-4rz8m"] Nov 25 12:23:43 crc kubenswrapper[4715]: I1125 12:23:43.165014 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl"] Nov 25 12:23:43 crc kubenswrapper[4715]: I1125 12:23:43.752309 4715 generic.go:334] "Generic (PLEG): container finished" podID="7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7" containerID="a4fd561c15a66b1f930af74f7b481476cd3c3658f14a06cd278d86fc5660d247" exitCode=0 Nov 25 12:23:43 crc kubenswrapper[4715]: I1125 12:23:43.752375 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" event={"ID":"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7","Type":"ContainerDied","Data":"a4fd561c15a66b1f930af74f7b481476cd3c3658f14a06cd278d86fc5660d247"} Nov 25 12:23:43 crc kubenswrapper[4715]: I1125 12:23:43.752618 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" event={"ID":"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7","Type":"ContainerStarted","Data":"cc174f50e7eaa31aa34777f4069e6fd55f3e4f81ec3a7ef33e5652fe4fbfd4e0"} Nov 25 12:23:43 crc kubenswrapper[4715]: I1125 12:23:43.753797 4715 generic.go:334] "Generic (PLEG): container finished" podID="d2d64e67-9d9c-46c1-855e-05f6343e223e" containerID="cca203ce6481b0cc50de66c6e121c614082f0d4fe3a52beeecd793596406a05a" exitCode=0 Nov 25 12:23:43 crc kubenswrapper[4715]: I1125 12:23:43.753837 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-4rz8m" event={"ID":"d2d64e67-9d9c-46c1-855e-05f6343e223e","Type":"ContainerDied","Data":"cca203ce6481b0cc50de66c6e121c614082f0d4fe3a52beeecd793596406a05a"} Nov 25 12:23:43 crc kubenswrapper[4715]: I1125 12:23:43.753886 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-4rz8m" event={"ID":"d2d64e67-9d9c-46c1-855e-05f6343e223e","Type":"ContainerStarted","Data":"8aacaf74aa1b6aed08057f1f7fc26ea834b9e693b4dd0acba2958370e210845f"} Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.053138 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.060295 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.111878 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-kvjlc"] Nov 25 12:23:45 crc kubenswrapper[4715]: E1125 12:23:45.112236 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7" containerName="mariadb-account-create-update" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.112261 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7" containerName="mariadb-account-create-update" Nov 25 12:23:45 crc kubenswrapper[4715]: E1125 12:23:45.112283 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d64e67-9d9c-46c1-855e-05f6343e223e" containerName="mariadb-database-create" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.112292 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d64e67-9d9c-46c1-855e-05f6343e223e" containerName="mariadb-database-create" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.112468 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7" containerName="mariadb-account-create-update" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.112488 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d64e67-9d9c-46c1-855e-05f6343e223e" containerName="mariadb-database-create" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.113066 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-kvjlc" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.116716 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-tfhg6" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.119732 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-kvjlc"] Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.157430 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2z5jm\" (UniqueName: \"kubernetes.io/projected/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-kube-api-access-2z5jm\") pod \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\" (UID: \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\") " Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.157724 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2d64e67-9d9c-46c1-855e-05f6343e223e-operator-scripts\") pod \"d2d64e67-9d9c-46c1-855e-05f6343e223e\" (UID: \"d2d64e67-9d9c-46c1-855e-05f6343e223e\") " Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.157862 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j4pc\" (UniqueName: \"kubernetes.io/projected/d2d64e67-9d9c-46c1-855e-05f6343e223e-kube-api-access-5j4pc\") pod \"d2d64e67-9d9c-46c1-855e-05f6343e223e\" (UID: \"d2d64e67-9d9c-46c1-855e-05f6343e223e\") " Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.157992 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-operator-scripts\") pod \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\" (UID: \"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7\") " Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.158209 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhmb4\" (UniqueName: \"kubernetes.io/projected/f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a-kube-api-access-mhmb4\") pod \"horizon-operator-index-kvjlc\" (UID: \"f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a\") " pod="openstack-operators/horizon-operator-index-kvjlc" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.158276 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2d64e67-9d9c-46c1-855e-05f6343e223e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2d64e67-9d9c-46c1-855e-05f6343e223e" (UID: "d2d64e67-9d9c-46c1-855e-05f6343e223e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.158608 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7" (UID: "7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.158624 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2d64e67-9d9c-46c1-855e-05f6343e223e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.163017 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-kube-api-access-2z5jm" (OuterVolumeSpecName: "kube-api-access-2z5jm") pod "7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7" (UID: "7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7"). InnerVolumeSpecName "kube-api-access-2z5jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.163620 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d64e67-9d9c-46c1-855e-05f6343e223e-kube-api-access-5j4pc" (OuterVolumeSpecName: "kube-api-access-5j4pc") pod "d2d64e67-9d9c-46c1-855e-05f6343e223e" (UID: "d2d64e67-9d9c-46c1-855e-05f6343e223e"). InnerVolumeSpecName "kube-api-access-5j4pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.259629 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhmb4\" (UniqueName: \"kubernetes.io/projected/f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a-kube-api-access-mhmb4\") pod \"horizon-operator-index-kvjlc\" (UID: \"f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a\") " pod="openstack-operators/horizon-operator-index-kvjlc" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.259756 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2z5jm\" (UniqueName: \"kubernetes.io/projected/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-kube-api-access-2z5jm\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.259768 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j4pc\" (UniqueName: \"kubernetes.io/projected/d2d64e67-9d9c-46c1-855e-05f6343e223e-kube-api-access-5j4pc\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.259778 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.275898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhmb4\" (UniqueName: \"kubernetes.io/projected/f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a-kube-api-access-mhmb4\") pod \"horizon-operator-index-kvjlc\" (UID: \"f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a\") " pod="openstack-operators/horizon-operator-index-kvjlc" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.432416 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-kvjlc" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.768897 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.768951 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl" event={"ID":"7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7","Type":"ContainerDied","Data":"cc174f50e7eaa31aa34777f4069e6fd55f3e4f81ec3a7ef33e5652fe4fbfd4e0"} Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.769258 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc174f50e7eaa31aa34777f4069e6fd55f3e4f81ec3a7ef33e5652fe4fbfd4e0" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.771473 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-4rz8m" event={"ID":"d2d64e67-9d9c-46c1-855e-05f6343e223e","Type":"ContainerDied","Data":"8aacaf74aa1b6aed08057f1f7fc26ea834b9e693b4dd0acba2958370e210845f"} Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.771510 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8aacaf74aa1b6aed08057f1f7fc26ea834b9e693b4dd0acba2958370e210845f" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.771530 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-4rz8m" Nov 25 12:23:45 crc kubenswrapper[4715]: I1125 12:23:45.830439 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-kvjlc"] Nov 25 12:23:45 crc kubenswrapper[4715]: W1125 12:23:45.833267 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf775cbfd_47ea_4fd8_9e1d_8a9e3ffbad0a.slice/crio-949ddf205d224d064fce4d5e2be736436770a377d0f0af9ee3e7bf0bc92e8181 WatchSource:0}: Error finding container 949ddf205d224d064fce4d5e2be736436770a377d0f0af9ee3e7bf0bc92e8181: Status 404 returned error can't find the container with id 949ddf205d224d064fce4d5e2be736436770a377d0f0af9ee3e7bf0bc92e8181 Nov 25 12:23:46 crc kubenswrapper[4715]: I1125 12:23:46.789367 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-kvjlc" event={"ID":"f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a","Type":"ContainerStarted","Data":"949ddf205d224d064fce4d5e2be736436770a377d0f0af9ee3e7bf0bc92e8181"} Nov 25 12:23:47 crc kubenswrapper[4715]: I1125 12:23:47.800689 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-kvjlc" event={"ID":"f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a","Type":"ContainerStarted","Data":"6328e87b4c704683757e4b3171ac16539909e17960d5939e5bc03cf8aee38fae"} Nov 25 12:23:47 crc kubenswrapper[4715]: I1125 12:23:47.828564 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-kvjlc" podStartSLOduration=1.712176628 podStartE2EDuration="2.828531103s" podCreationTimestamp="2025-11-25 12:23:45 +0000 UTC" firstStartedPulling="2025-11-25 12:23:45.835412095 +0000 UTC m=+876.342915116" lastFinishedPulling="2025-11-25 12:23:46.95176656 +0000 UTC m=+877.459269591" observedRunningTime="2025-11-25 12:23:47.82122693 +0000 UTC m=+878.328729991" watchObservedRunningTime="2025-11-25 12:23:47.828531103 +0000 UTC m=+878.336034164" Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.113652 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-bwlcw"] Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.115071 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-bwlcw" Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.119604 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-9pjhk" Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.131404 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-bwlcw"] Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.209682 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcvnb\" (UniqueName: \"kubernetes.io/projected/91b265ae-d562-4ebd-a3cc-57b86133d5c8-kube-api-access-zcvnb\") pod \"swift-operator-index-bwlcw\" (UID: \"91b265ae-d562-4ebd-a3cc-57b86133d5c8\") " pod="openstack-operators/swift-operator-index-bwlcw" Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.311202 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcvnb\" (UniqueName: \"kubernetes.io/projected/91b265ae-d562-4ebd-a3cc-57b86133d5c8-kube-api-access-zcvnb\") pod \"swift-operator-index-bwlcw\" (UID: \"91b265ae-d562-4ebd-a3cc-57b86133d5c8\") " pod="openstack-operators/swift-operator-index-bwlcw" Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.328579 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcvnb\" (UniqueName: \"kubernetes.io/projected/91b265ae-d562-4ebd-a3cc-57b86133d5c8-kube-api-access-zcvnb\") pod \"swift-operator-index-bwlcw\" (UID: \"91b265ae-d562-4ebd-a3cc-57b86133d5c8\") " pod="openstack-operators/swift-operator-index-bwlcw" Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.453489 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-bwlcw" Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.860096 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 25 12:23:48 crc kubenswrapper[4715]: I1125 12:23:48.885798 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-bwlcw"] Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.304585 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-hspqn"] Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.306065 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.307847 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.308232 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.308313 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.308313 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-gvsxs" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.315694 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-hspqn"] Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.424555 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th7sh\" (UniqueName: \"kubernetes.io/projected/2970fdeb-daab-4cdd-a29c-018c630df101-kube-api-access-th7sh\") pod \"keystone-db-sync-hspqn\" (UID: \"2970fdeb-daab-4cdd-a29c-018c630df101\") " pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.424704 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2970fdeb-daab-4cdd-a29c-018c630df101-config-data\") pod \"keystone-db-sync-hspqn\" (UID: \"2970fdeb-daab-4cdd-a29c-018c630df101\") " pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.526520 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th7sh\" (UniqueName: \"kubernetes.io/projected/2970fdeb-daab-4cdd-a29c-018c630df101-kube-api-access-th7sh\") pod \"keystone-db-sync-hspqn\" (UID: \"2970fdeb-daab-4cdd-a29c-018c630df101\") " pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.526646 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2970fdeb-daab-4cdd-a29c-018c630df101-config-data\") pod \"keystone-db-sync-hspqn\" (UID: \"2970fdeb-daab-4cdd-a29c-018c630df101\") " pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.532216 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2970fdeb-daab-4cdd-a29c-018c630df101-config-data\") pod \"keystone-db-sync-hspqn\" (UID: \"2970fdeb-daab-4cdd-a29c-018c630df101\") " pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.542079 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th7sh\" (UniqueName: \"kubernetes.io/projected/2970fdeb-daab-4cdd-a29c-018c630df101-kube-api-access-th7sh\") pod \"keystone-db-sync-hspqn\" (UID: \"2970fdeb-daab-4cdd-a29c-018c630df101\") " pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.624455 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:23:49 crc kubenswrapper[4715]: I1125 12:23:49.815809 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-bwlcw" event={"ID":"91b265ae-d562-4ebd-a3cc-57b86133d5c8","Type":"ContainerStarted","Data":"2fa0cc7bcf2e574e7a3e915e1db87d6b5bdef3ccbf63742cd207eb04663f3c5a"} Nov 25 12:23:50 crc kubenswrapper[4715]: I1125 12:23:50.040419 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-hspqn"] Nov 25 12:23:50 crc kubenswrapper[4715]: W1125 12:23:50.046884 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2970fdeb_daab_4cdd_a29c_018c630df101.slice/crio-1816d220be08d1a5a4a8f886b7f758ac4d24234c93661c0e37ce8382e902bbb2 WatchSource:0}: Error finding container 1816d220be08d1a5a4a8f886b7f758ac4d24234c93661c0e37ce8382e902bbb2: Status 404 returned error can't find the container with id 1816d220be08d1a5a4a8f886b7f758ac4d24234c93661c0e37ce8382e902bbb2 Nov 25 12:23:50 crc kubenswrapper[4715]: I1125 12:23:50.822590 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-hspqn" event={"ID":"2970fdeb-daab-4cdd-a29c-018c630df101","Type":"ContainerStarted","Data":"1816d220be08d1a5a4a8f886b7f758ac4d24234c93661c0e37ce8382e902bbb2"} Nov 25 12:23:51 crc kubenswrapper[4715]: I1125 12:23:51.507910 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-kvjlc"] Nov 25 12:23:51 crc kubenswrapper[4715]: I1125 12:23:51.509328 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/horizon-operator-index-kvjlc" podUID="f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a" containerName="registry-server" containerID="cri-o://6328e87b4c704683757e4b3171ac16539909e17960d5939e5bc03cf8aee38fae" gracePeriod=2 Nov 25 12:23:51 crc kubenswrapper[4715]: E1125 12:23:51.558818 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf775cbfd_47ea_4fd8_9e1d_8a9e3ffbad0a.slice/crio-6328e87b4c704683757e4b3171ac16539909e17960d5939e5bc03cf8aee38fae.scope\": RecentStats: unable to find data in memory cache]" Nov 25 12:23:51 crc kubenswrapper[4715]: I1125 12:23:51.829637 4715 generic.go:334] "Generic (PLEG): container finished" podID="f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a" containerID="6328e87b4c704683757e4b3171ac16539909e17960d5939e5bc03cf8aee38fae" exitCode=0 Nov 25 12:23:51 crc kubenswrapper[4715]: I1125 12:23:51.829707 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-kvjlc" event={"ID":"f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a","Type":"ContainerDied","Data":"6328e87b4c704683757e4b3171ac16539909e17960d5939e5bc03cf8aee38fae"} Nov 25 12:23:51 crc kubenswrapper[4715]: I1125 12:23:51.831255 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-bwlcw" event={"ID":"91b265ae-d562-4ebd-a3cc-57b86133d5c8","Type":"ContainerStarted","Data":"bd8d6d53fdcfd6a1050a13f9f965268aa6d07ddecbafac787a64d999e6181587"} Nov 25 12:23:51 crc kubenswrapper[4715]: I1125 12:23:51.847617 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-bwlcw" podStartSLOduration=1.684258772 podStartE2EDuration="3.847596074s" podCreationTimestamp="2025-11-25 12:23:48 +0000 UTC" firstStartedPulling="2025-11-25 12:23:48.898562247 +0000 UTC m=+879.406065268" lastFinishedPulling="2025-11-25 12:23:51.061899549 +0000 UTC m=+881.569402570" observedRunningTime="2025-11-25 12:23:51.844355974 +0000 UTC m=+882.351859005" watchObservedRunningTime="2025-11-25 12:23:51.847596074 +0000 UTC m=+882.355099095" Nov 25 12:23:51 crc kubenswrapper[4715]: I1125 12:23:51.882291 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-kvjlc" Nov 25 12:23:51 crc kubenswrapper[4715]: I1125 12:23:51.966885 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhmb4\" (UniqueName: \"kubernetes.io/projected/f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a-kube-api-access-mhmb4\") pod \"f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a\" (UID: \"f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a\") " Nov 25 12:23:51 crc kubenswrapper[4715]: I1125 12:23:51.972545 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a-kube-api-access-mhmb4" (OuterVolumeSpecName: "kube-api-access-mhmb4") pod "f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a" (UID: "f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a"). InnerVolumeSpecName "kube-api-access-mhmb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.068346 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhmb4\" (UniqueName: \"kubernetes.io/projected/f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a-kube-api-access-mhmb4\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.312752 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-67nk8"] Nov 25 12:23:52 crc kubenswrapper[4715]: E1125 12:23:52.313270 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a" containerName="registry-server" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.313354 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a" containerName="registry-server" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.313542 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a" containerName="registry-server" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.314060 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-67nk8" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.322037 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-67nk8"] Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.371839 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljp7g\" (UniqueName: \"kubernetes.io/projected/ac6f7b65-527d-4d7c-9c36-e2144540fc4c-kube-api-access-ljp7g\") pod \"horizon-operator-index-67nk8\" (UID: \"ac6f7b65-527d-4d7c-9c36-e2144540fc4c\") " pod="openstack-operators/horizon-operator-index-67nk8" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.473363 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljp7g\" (UniqueName: \"kubernetes.io/projected/ac6f7b65-527d-4d7c-9c36-e2144540fc4c-kube-api-access-ljp7g\") pod \"horizon-operator-index-67nk8\" (UID: \"ac6f7b65-527d-4d7c-9c36-e2144540fc4c\") " pod="openstack-operators/horizon-operator-index-67nk8" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.489979 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljp7g\" (UniqueName: \"kubernetes.io/projected/ac6f7b65-527d-4d7c-9c36-e2144540fc4c-kube-api-access-ljp7g\") pod \"horizon-operator-index-67nk8\" (UID: \"ac6f7b65-527d-4d7c-9c36-e2144540fc4c\") " pod="openstack-operators/horizon-operator-index-67nk8" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.632070 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-67nk8" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.844452 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-kvjlc" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.844582 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-kvjlc" event={"ID":"f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a","Type":"ContainerDied","Data":"949ddf205d224d064fce4d5e2be736436770a377d0f0af9ee3e7bf0bc92e8181"} Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.844632 4715 scope.go:117] "RemoveContainer" containerID="6328e87b4c704683757e4b3171ac16539909e17960d5939e5bc03cf8aee38fae" Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.867224 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-kvjlc"] Nov 25 12:23:52 crc kubenswrapper[4715]: I1125 12:23:52.872039 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/horizon-operator-index-kvjlc"] Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.102791 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-bwlcw"] Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.103313 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-index-bwlcw" podUID="91b265ae-d562-4ebd-a3cc-57b86133d5c8" containerName="registry-server" containerID="cri-o://bd8d6d53fdcfd6a1050a13f9f965268aa6d07ddecbafac787a64d999e6181587" gracePeriod=2 Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.703978 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a" path="/var/lib/kubelet/pods/f775cbfd-47ea-4fd8-9e1d-8a9e3ffbad0a/volumes" Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.704462 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-rhhbn"] Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.705231 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-rhhbn" Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.709981 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-rhhbn"] Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.808940 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b77qx\" (UniqueName: \"kubernetes.io/projected/3ab577a0-b9fa-4213-9513-648dcb1ff3fd-kube-api-access-b77qx\") pod \"swift-operator-index-rhhbn\" (UID: \"3ab577a0-b9fa-4213-9513-648dcb1ff3fd\") " pod="openstack-operators/swift-operator-index-rhhbn" Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.859461 4715 generic.go:334] "Generic (PLEG): container finished" podID="91b265ae-d562-4ebd-a3cc-57b86133d5c8" containerID="bd8d6d53fdcfd6a1050a13f9f965268aa6d07ddecbafac787a64d999e6181587" exitCode=0 Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.859510 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-bwlcw" event={"ID":"91b265ae-d562-4ebd-a3cc-57b86133d5c8","Type":"ContainerDied","Data":"bd8d6d53fdcfd6a1050a13f9f965268aa6d07ddecbafac787a64d999e6181587"} Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.909987 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b77qx\" (UniqueName: \"kubernetes.io/projected/3ab577a0-b9fa-4213-9513-648dcb1ff3fd-kube-api-access-b77qx\") pod \"swift-operator-index-rhhbn\" (UID: \"3ab577a0-b9fa-4213-9513-648dcb1ff3fd\") " pod="openstack-operators/swift-operator-index-rhhbn" Nov 25 12:23:54 crc kubenswrapper[4715]: I1125 12:23:54.939313 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b77qx\" (UniqueName: \"kubernetes.io/projected/3ab577a0-b9fa-4213-9513-648dcb1ff3fd-kube-api-access-b77qx\") pod \"swift-operator-index-rhhbn\" (UID: \"3ab577a0-b9fa-4213-9513-648dcb1ff3fd\") " pod="openstack-operators/swift-operator-index-rhhbn" Nov 25 12:23:55 crc kubenswrapper[4715]: I1125 12:23:55.020625 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-rhhbn" Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.452851 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-bwlcw" Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.531539 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcvnb\" (UniqueName: \"kubernetes.io/projected/91b265ae-d562-4ebd-a3cc-57b86133d5c8-kube-api-access-zcvnb\") pod \"91b265ae-d562-4ebd-a3cc-57b86133d5c8\" (UID: \"91b265ae-d562-4ebd-a3cc-57b86133d5c8\") " Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.539596 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91b265ae-d562-4ebd-a3cc-57b86133d5c8-kube-api-access-zcvnb" (OuterVolumeSpecName: "kube-api-access-zcvnb") pod "91b265ae-d562-4ebd-a3cc-57b86133d5c8" (UID: "91b265ae-d562-4ebd-a3cc-57b86133d5c8"). InnerVolumeSpecName "kube-api-access-zcvnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.632894 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcvnb\" (UniqueName: \"kubernetes.io/projected/91b265ae-d562-4ebd-a3cc-57b86133d5c8-kube-api-access-zcvnb\") on node \"crc\" DevicePath \"\"" Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.658839 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-67nk8"] Nov 25 12:23:56 crc kubenswrapper[4715]: W1125 12:23:56.664460 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac6f7b65_527d_4d7c_9c36_e2144540fc4c.slice/crio-363618638cb5f261223df870f2cdfee40f0c37024de64c981301967cd9bc8f33 WatchSource:0}: Error finding container 363618638cb5f261223df870f2cdfee40f0c37024de64c981301967cd9bc8f33: Status 404 returned error can't find the container with id 363618638cb5f261223df870f2cdfee40f0c37024de64c981301967cd9bc8f33 Nov 25 12:23:56 crc kubenswrapper[4715]: W1125 12:23:56.745377 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ab577a0_b9fa_4213_9513_648dcb1ff3fd.slice/crio-30bf39c8d2b7f1f1833299597d5b87ad80eaefd4c34d11ef5b133a0fa8faabeb WatchSource:0}: Error finding container 30bf39c8d2b7f1f1833299597d5b87ad80eaefd4c34d11ef5b133a0fa8faabeb: Status 404 returned error can't find the container with id 30bf39c8d2b7f1f1833299597d5b87ad80eaefd4c34d11ef5b133a0fa8faabeb Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.745489 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-rhhbn"] Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.874284 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-hspqn" event={"ID":"2970fdeb-daab-4cdd-a29c-018c630df101","Type":"ContainerStarted","Data":"2dc1c42c543a4346d8fe38356d143fb395d96a7534cbb33e570d56f74b758b39"} Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.876149 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-67nk8" event={"ID":"ac6f7b65-527d-4d7c-9c36-e2144540fc4c","Type":"ContainerStarted","Data":"363618638cb5f261223df870f2cdfee40f0c37024de64c981301967cd9bc8f33"} Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.879680 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-bwlcw" event={"ID":"91b265ae-d562-4ebd-a3cc-57b86133d5c8","Type":"ContainerDied","Data":"2fa0cc7bcf2e574e7a3e915e1db87d6b5bdef3ccbf63742cd207eb04663f3c5a"} Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.879740 4715 scope.go:117] "RemoveContainer" containerID="bd8d6d53fdcfd6a1050a13f9f965268aa6d07ddecbafac787a64d999e6181587" Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.879895 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-bwlcw" Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.882360 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-rhhbn" event={"ID":"3ab577a0-b9fa-4213-9513-648dcb1ff3fd","Type":"ContainerStarted","Data":"30bf39c8d2b7f1f1833299597d5b87ad80eaefd4c34d11ef5b133a0fa8faabeb"} Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.916284 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-hspqn" podStartSLOduration=1.662732275 podStartE2EDuration="7.916238275s" podCreationTimestamp="2025-11-25 12:23:49 +0000 UTC" firstStartedPulling="2025-11-25 12:23:50.049201818 +0000 UTC m=+880.556704839" lastFinishedPulling="2025-11-25 12:23:56.302707818 +0000 UTC m=+886.810210839" observedRunningTime="2025-11-25 12:23:56.896242602 +0000 UTC m=+887.403745703" watchObservedRunningTime="2025-11-25 12:23:56.916238275 +0000 UTC m=+887.423741306" Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.925786 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-bwlcw"] Nov 25 12:23:56 crc kubenswrapper[4715]: I1125 12:23:56.932333 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-index-bwlcw"] Nov 25 12:23:57 crc kubenswrapper[4715]: I1125 12:23:57.892012 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-rhhbn" event={"ID":"3ab577a0-b9fa-4213-9513-648dcb1ff3fd","Type":"ContainerStarted","Data":"84ae1e748bfd0ae11490b8ff3a257996bfa9ff514315e8656cf1c160a1e44f2e"} Nov 25 12:23:57 crc kubenswrapper[4715]: I1125 12:23:57.893996 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-67nk8" event={"ID":"ac6f7b65-527d-4d7c-9c36-e2144540fc4c","Type":"ContainerStarted","Data":"ffd5165aace2fb3691e9ac459c3c40a7354068510ef7f0b52b1105668e7f7ac6"} Nov 25 12:23:57 crc kubenswrapper[4715]: I1125 12:23:57.914668 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-rhhbn" podStartSLOduration=3.358827708 podStartE2EDuration="3.91464805s" podCreationTimestamp="2025-11-25 12:23:54 +0000 UTC" firstStartedPulling="2025-11-25 12:23:56.749137156 +0000 UTC m=+887.256640177" lastFinishedPulling="2025-11-25 12:23:57.304957488 +0000 UTC m=+887.812460519" observedRunningTime="2025-11-25 12:23:57.910265538 +0000 UTC m=+888.417768579" watchObservedRunningTime="2025-11-25 12:23:57.91464805 +0000 UTC m=+888.422151081" Nov 25 12:23:57 crc kubenswrapper[4715]: I1125 12:23:57.930289 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-67nk8" podStartSLOduration=5.420990125 podStartE2EDuration="5.930274312s" podCreationTimestamp="2025-11-25 12:23:52 +0000 UTC" firstStartedPulling="2025-11-25 12:23:56.671247053 +0000 UTC m=+887.178750094" lastFinishedPulling="2025-11-25 12:23:57.18053125 +0000 UTC m=+887.688034281" observedRunningTime="2025-11-25 12:23:57.927723621 +0000 UTC m=+888.435226642" watchObservedRunningTime="2025-11-25 12:23:57.930274312 +0000 UTC m=+888.437777333" Nov 25 12:23:58 crc kubenswrapper[4715]: I1125 12:23:58.701306 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91b265ae-d562-4ebd-a3cc-57b86133d5c8" path="/var/lib/kubelet/pods/91b265ae-d562-4ebd-a3cc-57b86133d5c8/volumes" Nov 25 12:24:00 crc kubenswrapper[4715]: I1125 12:24:00.918315 4715 generic.go:334] "Generic (PLEG): container finished" podID="2970fdeb-daab-4cdd-a29c-018c630df101" containerID="2dc1c42c543a4346d8fe38356d143fb395d96a7534cbb33e570d56f74b758b39" exitCode=0 Nov 25 12:24:00 crc kubenswrapper[4715]: I1125 12:24:00.918427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-hspqn" event={"ID":"2970fdeb-daab-4cdd-a29c-018c630df101","Type":"ContainerDied","Data":"2dc1c42c543a4346d8fe38356d143fb395d96a7534cbb33e570d56f74b758b39"} Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.191840 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.310997 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th7sh\" (UniqueName: \"kubernetes.io/projected/2970fdeb-daab-4cdd-a29c-018c630df101-kube-api-access-th7sh\") pod \"2970fdeb-daab-4cdd-a29c-018c630df101\" (UID: \"2970fdeb-daab-4cdd-a29c-018c630df101\") " Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.311099 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2970fdeb-daab-4cdd-a29c-018c630df101-config-data\") pod \"2970fdeb-daab-4cdd-a29c-018c630df101\" (UID: \"2970fdeb-daab-4cdd-a29c-018c630df101\") " Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.316219 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2970fdeb-daab-4cdd-a29c-018c630df101-kube-api-access-th7sh" (OuterVolumeSpecName: "kube-api-access-th7sh") pod "2970fdeb-daab-4cdd-a29c-018c630df101" (UID: "2970fdeb-daab-4cdd-a29c-018c630df101"). InnerVolumeSpecName "kube-api-access-th7sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.343730 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2970fdeb-daab-4cdd-a29c-018c630df101-config-data" (OuterVolumeSpecName: "config-data") pod "2970fdeb-daab-4cdd-a29c-018c630df101" (UID: "2970fdeb-daab-4cdd-a29c-018c630df101"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.412400 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th7sh\" (UniqueName: \"kubernetes.io/projected/2970fdeb-daab-4cdd-a29c-018c630df101-kube-api-access-th7sh\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.412431 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2970fdeb-daab-4cdd-a29c-018c630df101-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.633315 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-67nk8" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.633356 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-67nk8" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.682489 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-67nk8" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.930550 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-hspqn" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.930545 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-hspqn" event={"ID":"2970fdeb-daab-4cdd-a29c-018c630df101","Type":"ContainerDied","Data":"1816d220be08d1a5a4a8f886b7f758ac4d24234c93661c0e37ce8382e902bbb2"} Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.930657 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1816d220be08d1a5a4a8f886b7f758ac4d24234c93661c0e37ce8382e902bbb2" Nov 25 12:24:02 crc kubenswrapper[4715]: I1125 12:24:02.959839 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-67nk8" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.119073 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-hw97n"] Nov 25 12:24:03 crc kubenswrapper[4715]: E1125 12:24:03.119394 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2970fdeb-daab-4cdd-a29c-018c630df101" containerName="keystone-db-sync" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.119418 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2970fdeb-daab-4cdd-a29c-018c630df101" containerName="keystone-db-sync" Nov 25 12:24:03 crc kubenswrapper[4715]: E1125 12:24:03.119456 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91b265ae-d562-4ebd-a3cc-57b86133d5c8" containerName="registry-server" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.119466 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="91b265ae-d562-4ebd-a3cc-57b86133d5c8" containerName="registry-server" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.119612 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="91b265ae-d562-4ebd-a3cc-57b86133d5c8" containerName="registry-server" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.119642 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2970fdeb-daab-4cdd-a29c-018c630df101" containerName="keystone-db-sync" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.120159 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.123084 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-gvsxs" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.123298 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.123408 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.123505 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.123801 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.132172 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-hw97n"] Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.222652 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-scripts\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.222748 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-fernet-keys\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.222783 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-credential-keys\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.222859 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsd8r\" (UniqueName: \"kubernetes.io/projected/77036b2e-4608-4b32-bf8c-27099d539710-kube-api-access-vsd8r\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.223032 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-config-data\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.324665 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-scripts\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.324730 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-fernet-keys\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.324760 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-credential-keys\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.324786 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsd8r\" (UniqueName: \"kubernetes.io/projected/77036b2e-4608-4b32-bf8c-27099d539710-kube-api-access-vsd8r\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.324843 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-config-data\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.329101 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-fernet-keys\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.329141 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-credential-keys\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.329639 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-scripts\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.329898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-config-data\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.340768 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsd8r\" (UniqueName: \"kubernetes.io/projected/77036b2e-4608-4b32-bf8c-27099d539710-kube-api-access-vsd8r\") pod \"keystone-bootstrap-hw97n\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.434784 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.836862 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-hw97n"] Nov 25 12:24:03 crc kubenswrapper[4715]: I1125 12:24:03.938146 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-hw97n" event={"ID":"77036b2e-4608-4b32-bf8c-27099d539710","Type":"ContainerStarted","Data":"e704ce085203a4ee0910bbbede90863fa29aef0bf30d20e2c9a82938c3307474"} Nov 25 12:24:04 crc kubenswrapper[4715]: I1125 12:24:04.953541 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-hw97n" event={"ID":"77036b2e-4608-4b32-bf8c-27099d539710","Type":"ContainerStarted","Data":"5b9c3eb7e2c24f222fd8a35655b3f27a95f3fe258f9eb0b63762658998077f01"} Nov 25 12:24:04 crc kubenswrapper[4715]: I1125 12:24:04.970591 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-hw97n" podStartSLOduration=1.9705774649999999 podStartE2EDuration="1.970577465s" podCreationTimestamp="2025-11-25 12:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:24:04.968296582 +0000 UTC m=+895.475799603" watchObservedRunningTime="2025-11-25 12:24:04.970577465 +0000 UTC m=+895.478080486" Nov 25 12:24:05 crc kubenswrapper[4715]: I1125 12:24:05.021638 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-rhhbn" Nov 25 12:24:05 crc kubenswrapper[4715]: I1125 12:24:05.021689 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-rhhbn" Nov 25 12:24:05 crc kubenswrapper[4715]: I1125 12:24:05.047629 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-rhhbn" Nov 25 12:24:05 crc kubenswrapper[4715]: I1125 12:24:05.992354 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-rhhbn" Nov 25 12:24:06 crc kubenswrapper[4715]: I1125 12:24:06.971899 4715 generic.go:334] "Generic (PLEG): container finished" podID="77036b2e-4608-4b32-bf8c-27099d539710" containerID="5b9c3eb7e2c24f222fd8a35655b3f27a95f3fe258f9eb0b63762658998077f01" exitCode=0 Nov 25 12:24:06 crc kubenswrapper[4715]: I1125 12:24:06.971997 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-hw97n" event={"ID":"77036b2e-4608-4b32-bf8c-27099d539710","Type":"ContainerDied","Data":"5b9c3eb7e2c24f222fd8a35655b3f27a95f3fe258f9eb0b63762658998077f01"} Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.238517 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.309130 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-credential-keys\") pod \"77036b2e-4608-4b32-bf8c-27099d539710\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.309199 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-fernet-keys\") pod \"77036b2e-4608-4b32-bf8c-27099d539710\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.309295 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-config-data\") pod \"77036b2e-4608-4b32-bf8c-27099d539710\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.309347 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-scripts\") pod \"77036b2e-4608-4b32-bf8c-27099d539710\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.309392 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsd8r\" (UniqueName: \"kubernetes.io/projected/77036b2e-4608-4b32-bf8c-27099d539710-kube-api-access-vsd8r\") pod \"77036b2e-4608-4b32-bf8c-27099d539710\" (UID: \"77036b2e-4608-4b32-bf8c-27099d539710\") " Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.314523 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "77036b2e-4608-4b32-bf8c-27099d539710" (UID: "77036b2e-4608-4b32-bf8c-27099d539710"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.315036 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-scripts" (OuterVolumeSpecName: "scripts") pod "77036b2e-4608-4b32-bf8c-27099d539710" (UID: "77036b2e-4608-4b32-bf8c-27099d539710"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.315114 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "77036b2e-4608-4b32-bf8c-27099d539710" (UID: "77036b2e-4608-4b32-bf8c-27099d539710"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.316070 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77036b2e-4608-4b32-bf8c-27099d539710-kube-api-access-vsd8r" (OuterVolumeSpecName: "kube-api-access-vsd8r") pod "77036b2e-4608-4b32-bf8c-27099d539710" (UID: "77036b2e-4608-4b32-bf8c-27099d539710"). InnerVolumeSpecName "kube-api-access-vsd8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.330500 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-config-data" (OuterVolumeSpecName: "config-data") pod "77036b2e-4608-4b32-bf8c-27099d539710" (UID: "77036b2e-4608-4b32-bf8c-27099d539710"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.411254 4715 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.411294 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.411306 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.411317 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77036b2e-4608-4b32-bf8c-27099d539710-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.411327 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsd8r\" (UniqueName: \"kubernetes.io/projected/77036b2e-4608-4b32-bf8c-27099d539710-kube-api-access-vsd8r\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.988984 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-hw97n" event={"ID":"77036b2e-4608-4b32-bf8c-27099d539710","Type":"ContainerDied","Data":"e704ce085203a4ee0910bbbede90863fa29aef0bf30d20e2c9a82938c3307474"} Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.989030 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e704ce085203a4ee0910bbbede90863fa29aef0bf30d20e2c9a82938c3307474" Nov 25 12:24:08 crc kubenswrapper[4715]: I1125 12:24:08.989089 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-hw97n" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.169909 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-7f7c6799b4-5dgxc"] Nov 25 12:24:09 crc kubenswrapper[4715]: E1125 12:24:09.170169 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77036b2e-4608-4b32-bf8c-27099d539710" containerName="keystone-bootstrap" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.170201 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="77036b2e-4608-4b32-bf8c-27099d539710" containerName="keystone-bootstrap" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.170316 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="77036b2e-4608-4b32-bf8c-27099d539710" containerName="keystone-bootstrap" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.170711 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.173100 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-gvsxs" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.173989 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.174238 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.175347 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.180897 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-7f7c6799b4-5dgxc"] Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.220131 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-scripts\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.220260 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-credential-keys\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.220296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-fernet-keys\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.220324 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-config-data\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.220403 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5nn9\" (UniqueName: \"kubernetes.io/projected/2b2280ed-0d22-4a77-8881-b91088f48347-kube-api-access-p5nn9\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.321794 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-fernet-keys\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.321850 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-config-data\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.322419 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5nn9\" (UniqueName: \"kubernetes.io/projected/2b2280ed-0d22-4a77-8881-b91088f48347-kube-api-access-p5nn9\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.322499 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-scripts\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.322556 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-credential-keys\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.325863 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-fernet-keys\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.325936 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-config-data\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.326263 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-credential-keys\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.326400 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b2280ed-0d22-4a77-8881-b91088f48347-scripts\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.341114 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5nn9\" (UniqueName: \"kubernetes.io/projected/2b2280ed-0d22-4a77-8881-b91088f48347-kube-api-access-p5nn9\") pod \"keystone-7f7c6799b4-5dgxc\" (UID: \"2b2280ed-0d22-4a77-8881-b91088f48347\") " pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.489199 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.518954 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k2crl"] Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.521027 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.531548 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2crl"] Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.626725 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsmw5\" (UniqueName: \"kubernetes.io/projected/cea6d8c6-58c6-4abb-b039-af291a90e8be-kube-api-access-rsmw5\") pod \"community-operators-k2crl\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.627017 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-catalog-content\") pod \"community-operators-k2crl\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.627073 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-utilities\") pod \"community-operators-k2crl\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.727928 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-utilities\") pod \"community-operators-k2crl\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.728016 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsmw5\" (UniqueName: \"kubernetes.io/projected/cea6d8c6-58c6-4abb-b039-af291a90e8be-kube-api-access-rsmw5\") pod \"community-operators-k2crl\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.728057 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-catalog-content\") pod \"community-operators-k2crl\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.728512 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-utilities\") pod \"community-operators-k2crl\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.728566 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-catalog-content\") pod \"community-operators-k2crl\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.759285 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsmw5\" (UniqueName: \"kubernetes.io/projected/cea6d8c6-58c6-4abb-b039-af291a90e8be-kube-api-access-rsmw5\") pod \"community-operators-k2crl\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.904512 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:09 crc kubenswrapper[4715]: I1125 12:24:09.958973 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-7f7c6799b4-5dgxc"] Nov 25 12:24:09 crc kubenswrapper[4715]: W1125 12:24:09.978548 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b2280ed_0d22_4a77_8881_b91088f48347.slice/crio-1b7b2f314442fd0fa448f3debf1955c1fefdf97903b85d1580bae772464043db WatchSource:0}: Error finding container 1b7b2f314442fd0fa448f3debf1955c1fefdf97903b85d1580bae772464043db: Status 404 returned error can't find the container with id 1b7b2f314442fd0fa448f3debf1955c1fefdf97903b85d1580bae772464043db Nov 25 12:24:10 crc kubenswrapper[4715]: I1125 12:24:10.002364 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" event={"ID":"2b2280ed-0d22-4a77-8881-b91088f48347","Type":"ContainerStarted","Data":"1b7b2f314442fd0fa448f3debf1955c1fefdf97903b85d1580bae772464043db"} Nov 25 12:24:10 crc kubenswrapper[4715]: I1125 12:24:10.364040 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2crl"] Nov 25 12:24:10 crc kubenswrapper[4715]: W1125 12:24:10.370761 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcea6d8c6_58c6_4abb_b039_af291a90e8be.slice/crio-47b0e1336a7953a6b8a05e6da7ffe9dd82653091a80038495336c257f7ab914a WatchSource:0}: Error finding container 47b0e1336a7953a6b8a05e6da7ffe9dd82653091a80038495336c257f7ab914a: Status 404 returned error can't find the container with id 47b0e1336a7953a6b8a05e6da7ffe9dd82653091a80038495336c257f7ab914a Nov 25 12:24:11 crc kubenswrapper[4715]: I1125 12:24:11.010311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" event={"ID":"2b2280ed-0d22-4a77-8881-b91088f48347","Type":"ContainerStarted","Data":"093991bb0631a2234aef367b66bd9f3a6c3d8a834b34ebfa3129d18200cf024d"} Nov 25 12:24:11 crc kubenswrapper[4715]: I1125 12:24:11.010853 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:11 crc kubenswrapper[4715]: I1125 12:24:11.011961 4715 generic.go:334] "Generic (PLEG): container finished" podID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerID="c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e" exitCode=0 Nov 25 12:24:11 crc kubenswrapper[4715]: I1125 12:24:11.012012 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2crl" event={"ID":"cea6d8c6-58c6-4abb-b039-af291a90e8be","Type":"ContainerDied","Data":"c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e"} Nov 25 12:24:11 crc kubenswrapper[4715]: I1125 12:24:11.012052 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2crl" event={"ID":"cea6d8c6-58c6-4abb-b039-af291a90e8be","Type":"ContainerStarted","Data":"47b0e1336a7953a6b8a05e6da7ffe9dd82653091a80038495336c257f7ab914a"} Nov 25 12:24:11 crc kubenswrapper[4715]: I1125 12:24:11.027749 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" podStartSLOduration=2.027726727 podStartE2EDuration="2.027726727s" podCreationTimestamp="2025-11-25 12:24:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:24:11.023631234 +0000 UTC m=+901.531134255" watchObservedRunningTime="2025-11-25 12:24:11.027726727 +0000 UTC m=+901.535229748" Nov 25 12:24:12 crc kubenswrapper[4715]: I1125 12:24:12.020213 4715 generic.go:334] "Generic (PLEG): container finished" podID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerID="5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28" exitCode=0 Nov 25 12:24:12 crc kubenswrapper[4715]: I1125 12:24:12.020288 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2crl" event={"ID":"cea6d8c6-58c6-4abb-b039-af291a90e8be","Type":"ContainerDied","Data":"5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28"} Nov 25 12:24:13 crc kubenswrapper[4715]: I1125 12:24:13.028366 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2crl" event={"ID":"cea6d8c6-58c6-4abb-b039-af291a90e8be","Type":"ContainerStarted","Data":"bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000"} Nov 25 12:24:13 crc kubenswrapper[4715]: I1125 12:24:13.048524 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k2crl" podStartSLOduration=2.563559467 podStartE2EDuration="4.048500618s" podCreationTimestamp="2025-11-25 12:24:09 +0000 UTC" firstStartedPulling="2025-11-25 12:24:11.013368751 +0000 UTC m=+901.520871772" lastFinishedPulling="2025-11-25 12:24:12.498309902 +0000 UTC m=+903.005812923" observedRunningTime="2025-11-25 12:24:13.044223801 +0000 UTC m=+903.551726842" watchObservedRunningTime="2025-11-25 12:24:13.048500618 +0000 UTC m=+903.556003639" Nov 25 12:24:19 crc kubenswrapper[4715]: I1125 12:24:19.904773 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:19 crc kubenswrapper[4715]: I1125 12:24:19.905409 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:19 crc kubenswrapper[4715]: I1125 12:24:19.948018 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.111751 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.558638 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj"] Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.560176 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.563672 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7xsfb" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.569744 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj"] Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.682803 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-bundle\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.682892 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-util\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.683120 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtsww\" (UniqueName: \"kubernetes.io/projected/731e898e-a432-4696-ab69-cc1e0263558e-kube-api-access-rtsww\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.784895 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-util\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.785066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtsww\" (UniqueName: \"kubernetes.io/projected/731e898e-a432-4696-ab69-cc1e0263558e-kube-api-access-rtsww\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.785298 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-bundle\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.785698 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-util\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.786023 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-bundle\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.808234 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtsww\" (UniqueName: \"kubernetes.io/projected/731e898e-a432-4696-ab69-cc1e0263558e-kube-api-access-rtsww\") pod \"aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:20 crc kubenswrapper[4715]: I1125 12:24:20.886728 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.346437 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj"] Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.537362 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f"] Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.538569 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.546068 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f"] Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.594475 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gw7b\" (UniqueName: \"kubernetes.io/projected/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-kube-api-access-8gw7b\") pod \"2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.594568 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-bundle\") pod \"2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.594626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-util\") pod \"2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.695645 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gw7b\" (UniqueName: \"kubernetes.io/projected/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-kube-api-access-8gw7b\") pod \"2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.695722 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-bundle\") pod \"2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.695771 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-util\") pod \"2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.696286 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-util\") pod \"2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.696330 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-bundle\") pod \"2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.713716 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gw7b\" (UniqueName: \"kubernetes.io/projected/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-kube-api-access-8gw7b\") pod \"2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:21 crc kubenswrapper[4715]: I1125 12:24:21.852068 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:22 crc kubenswrapper[4715]: I1125 12:24:22.059620 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f"] Nov 25 12:24:22 crc kubenswrapper[4715]: I1125 12:24:22.096463 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" event={"ID":"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b","Type":"ContainerStarted","Data":"bfbfa266386bc7dea9dea19f943e6b5da7a1212fc001940ef5bbcb0636562241"} Nov 25 12:24:22 crc kubenswrapper[4715]: I1125 12:24:22.100110 4715 generic.go:334] "Generic (PLEG): container finished" podID="731e898e-a432-4696-ab69-cc1e0263558e" containerID="1377eb6c556375864b76118f43a398d6373ee85e7d6ec5c18d3c545d448a9f74" exitCode=0 Nov 25 12:24:22 crc kubenswrapper[4715]: I1125 12:24:22.100300 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" event={"ID":"731e898e-a432-4696-ab69-cc1e0263558e","Type":"ContainerDied","Data":"1377eb6c556375864b76118f43a398d6373ee85e7d6ec5c18d3c545d448a9f74"} Nov 25 12:24:22 crc kubenswrapper[4715]: I1125 12:24:22.100336 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" event={"ID":"731e898e-a432-4696-ab69-cc1e0263558e","Type":"ContainerStarted","Data":"ee9eebe1220464a7b19cec8dec550b08c0b6a433ad658eadfaa7100656164958"} Nov 25 12:24:23 crc kubenswrapper[4715]: I1125 12:24:23.107541 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" event={"ID":"731e898e-a432-4696-ab69-cc1e0263558e","Type":"ContainerStarted","Data":"8f8c56d14d3e0d6553d168754d4da3a6c301880c1141a9d3827dbeecd2f45d64"} Nov 25 12:24:23 crc kubenswrapper[4715]: I1125 12:24:23.108927 4715 generic.go:334] "Generic (PLEG): container finished" podID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerID="46f27565780acd1ef5cbf0632f9f828e8a3370010016ed35c65da44cea037659" exitCode=0 Nov 25 12:24:23 crc kubenswrapper[4715]: I1125 12:24:23.108955 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" event={"ID":"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b","Type":"ContainerDied","Data":"46f27565780acd1ef5cbf0632f9f828e8a3370010016ed35c65da44cea037659"} Nov 25 12:24:24 crc kubenswrapper[4715]: I1125 12:24:24.115903 4715 generic.go:334] "Generic (PLEG): container finished" podID="731e898e-a432-4696-ab69-cc1e0263558e" containerID="8f8c56d14d3e0d6553d168754d4da3a6c301880c1141a9d3827dbeecd2f45d64" exitCode=0 Nov 25 12:24:24 crc kubenswrapper[4715]: I1125 12:24:24.115953 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" event={"ID":"731e898e-a432-4696-ab69-cc1e0263558e","Type":"ContainerDied","Data":"8f8c56d14d3e0d6553d168754d4da3a6c301880c1141a9d3827dbeecd2f45d64"} Nov 25 12:24:24 crc kubenswrapper[4715]: I1125 12:24:24.117797 4715 generic.go:334] "Generic (PLEG): container finished" podID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerID="fa6547c274e62ee3a132d798dc79259ad5b5734c8e3bbb92818865da7c3204ad" exitCode=0 Nov 25 12:24:24 crc kubenswrapper[4715]: I1125 12:24:24.117827 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" event={"ID":"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b","Type":"ContainerDied","Data":"fa6547c274e62ee3a132d798dc79259ad5b5734c8e3bbb92818865da7c3204ad"} Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.125974 4715 generic.go:334] "Generic (PLEG): container finished" podID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerID="34bfd4c15527fb6b3190cb14d575d5ae137641b0b2a7755793f8f691f3edf808" exitCode=0 Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.126026 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" event={"ID":"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b","Type":"ContainerDied","Data":"34bfd4c15527fb6b3190cb14d575d5ae137641b0b2a7755793f8f691f3edf808"} Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.128969 4715 generic.go:334] "Generic (PLEG): container finished" podID="731e898e-a432-4696-ab69-cc1e0263558e" containerID="898b46caceec465038dec2fab84bbe4ef59b5240d8b8615053f8804ef36c948c" exitCode=0 Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.129012 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" event={"ID":"731e898e-a432-4696-ab69-cc1e0263558e","Type":"ContainerDied","Data":"898b46caceec465038dec2fab84bbe4ef59b5240d8b8615053f8804ef36c948c"} Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.304327 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2crl"] Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.304552 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k2crl" podUID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerName="registry-server" containerID="cri-o://bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000" gracePeriod=2 Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.699766 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.754518 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsmw5\" (UniqueName: \"kubernetes.io/projected/cea6d8c6-58c6-4abb-b039-af291a90e8be-kube-api-access-rsmw5\") pod \"cea6d8c6-58c6-4abb-b039-af291a90e8be\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.754580 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-catalog-content\") pod \"cea6d8c6-58c6-4abb-b039-af291a90e8be\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.754626 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-utilities\") pod \"cea6d8c6-58c6-4abb-b039-af291a90e8be\" (UID: \"cea6d8c6-58c6-4abb-b039-af291a90e8be\") " Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.756729 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-utilities" (OuterVolumeSpecName: "utilities") pod "cea6d8c6-58c6-4abb-b039-af291a90e8be" (UID: "cea6d8c6-58c6-4abb-b039-af291a90e8be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.763635 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cea6d8c6-58c6-4abb-b039-af291a90e8be-kube-api-access-rsmw5" (OuterVolumeSpecName: "kube-api-access-rsmw5") pod "cea6d8c6-58c6-4abb-b039-af291a90e8be" (UID: "cea6d8c6-58c6-4abb-b039-af291a90e8be"). InnerVolumeSpecName "kube-api-access-rsmw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.807651 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cea6d8c6-58c6-4abb-b039-af291a90e8be" (UID: "cea6d8c6-58c6-4abb-b039-af291a90e8be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.856407 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsmw5\" (UniqueName: \"kubernetes.io/projected/cea6d8c6-58c6-4abb-b039-af291a90e8be-kube-api-access-rsmw5\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.856457 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:25 crc kubenswrapper[4715]: I1125 12:24:25.856468 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cea6d8c6-58c6-4abb-b039-af291a90e8be-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.144766 4715 generic.go:334] "Generic (PLEG): container finished" podID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerID="bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000" exitCode=0 Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.144846 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2crl" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.144894 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2crl" event={"ID":"cea6d8c6-58c6-4abb-b039-af291a90e8be","Type":"ContainerDied","Data":"bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000"} Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.144964 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2crl" event={"ID":"cea6d8c6-58c6-4abb-b039-af291a90e8be","Type":"ContainerDied","Data":"47b0e1336a7953a6b8a05e6da7ffe9dd82653091a80038495336c257f7ab914a"} Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.144992 4715 scope.go:117] "RemoveContainer" containerID="bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.171721 4715 scope.go:117] "RemoveContainer" containerID="5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.180267 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2crl"] Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.187218 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k2crl"] Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.203811 4715 scope.go:117] "RemoveContainer" containerID="c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.231434 4715 scope.go:117] "RemoveContainer" containerID="bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000" Nov 25 12:24:26 crc kubenswrapper[4715]: E1125 12:24:26.231836 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000\": container with ID starting with bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000 not found: ID does not exist" containerID="bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.231872 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000"} err="failed to get container status \"bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000\": rpc error: code = NotFound desc = could not find container \"bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000\": container with ID starting with bd28f86b5ea81240d7a7509c9fa47ade7bd69c672ea67a4df8340701032f2000 not found: ID does not exist" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.231898 4715 scope.go:117] "RemoveContainer" containerID="5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28" Nov 25 12:24:26 crc kubenswrapper[4715]: E1125 12:24:26.232176 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28\": container with ID starting with 5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28 not found: ID does not exist" containerID="5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.232221 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28"} err="failed to get container status \"5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28\": rpc error: code = NotFound desc = could not find container \"5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28\": container with ID starting with 5dc3b40a4c638fc213eff4d3be4c3cbbeebf82e5fa41f6ca79de19ec3046de28 not found: ID does not exist" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.232240 4715 scope.go:117] "RemoveContainer" containerID="c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e" Nov 25 12:24:26 crc kubenswrapper[4715]: E1125 12:24:26.232504 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e\": container with ID starting with c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e not found: ID does not exist" containerID="c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.232524 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e"} err="failed to get container status \"c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e\": rpc error: code = NotFound desc = could not find container \"c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e\": container with ID starting with c9888e688a192d8f7785a31554e3c5b10d8aa6068944da300cf857fa1c04d60e not found: ID does not exist" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.381538 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.386728 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.464084 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtsww\" (UniqueName: \"kubernetes.io/projected/731e898e-a432-4696-ab69-cc1e0263558e-kube-api-access-rtsww\") pod \"731e898e-a432-4696-ab69-cc1e0263558e\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.464336 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-bundle\") pod \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.464451 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-bundle\") pod \"731e898e-a432-4696-ab69-cc1e0263558e\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.464559 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-util\") pod \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.464663 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-util\") pod \"731e898e-a432-4696-ab69-cc1e0263558e\" (UID: \"731e898e-a432-4696-ab69-cc1e0263558e\") " Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.464725 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gw7b\" (UniqueName: \"kubernetes.io/projected/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-kube-api-access-8gw7b\") pod \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\" (UID: \"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b\") " Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.465290 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-bundle" (OuterVolumeSpecName: "bundle") pod "731e898e-a432-4696-ab69-cc1e0263558e" (UID: "731e898e-a432-4696-ab69-cc1e0263558e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.465464 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-bundle" (OuterVolumeSpecName: "bundle") pod "5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" (UID: "5d1aad28-aabe-40fb-8d6c-3d3643a6a52b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.468312 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/731e898e-a432-4696-ab69-cc1e0263558e-kube-api-access-rtsww" (OuterVolumeSpecName: "kube-api-access-rtsww") pod "731e898e-a432-4696-ab69-cc1e0263558e" (UID: "731e898e-a432-4696-ab69-cc1e0263558e"). InnerVolumeSpecName "kube-api-access-rtsww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.468436 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-kube-api-access-8gw7b" (OuterVolumeSpecName: "kube-api-access-8gw7b") pod "5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" (UID: "5d1aad28-aabe-40fb-8d6c-3d3643a6a52b"). InnerVolumeSpecName "kube-api-access-8gw7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.477236 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-util" (OuterVolumeSpecName: "util") pod "5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" (UID: "5d1aad28-aabe-40fb-8d6c-3d3643a6a52b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.477590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-util" (OuterVolumeSpecName: "util") pod "731e898e-a432-4696-ab69-cc1e0263558e" (UID: "731e898e-a432-4696-ab69-cc1e0263558e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.566902 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.566938 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.566947 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-util\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.566955 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/731e898e-a432-4696-ab69-cc1e0263558e-util\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.566965 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gw7b\" (UniqueName: \"kubernetes.io/projected/5d1aad28-aabe-40fb-8d6c-3d3643a6a52b-kube-api-access-8gw7b\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.566975 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtsww\" (UniqueName: \"kubernetes.io/projected/731e898e-a432-4696-ab69-cc1e0263558e-kube-api-access-rtsww\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:26 crc kubenswrapper[4715]: I1125 12:24:26.701836 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cea6d8c6-58c6-4abb-b039-af291a90e8be" path="/var/lib/kubelet/pods/cea6d8c6-58c6-4abb-b039-af291a90e8be/volumes" Nov 25 12:24:27 crc kubenswrapper[4715]: I1125 12:24:27.158291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" event={"ID":"5d1aad28-aabe-40fb-8d6c-3d3643a6a52b","Type":"ContainerDied","Data":"bfbfa266386bc7dea9dea19f943e6b5da7a1212fc001940ef5bbcb0636562241"} Nov 25 12:24:27 crc kubenswrapper[4715]: I1125 12:24:27.158810 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfbfa266386bc7dea9dea19f943e6b5da7a1212fc001940ef5bbcb0636562241" Nov 25 12:24:27 crc kubenswrapper[4715]: I1125 12:24:27.158419 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f" Nov 25 12:24:27 crc kubenswrapper[4715]: I1125 12:24:27.165808 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" event={"ID":"731e898e-a432-4696-ab69-cc1e0263558e","Type":"ContainerDied","Data":"ee9eebe1220464a7b19cec8dec550b08c0b6a433ad658eadfaa7100656164958"} Nov 25 12:24:27 crc kubenswrapper[4715]: I1125 12:24:27.165862 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee9eebe1220464a7b19cec8dec550b08c0b6a433ad658eadfaa7100656164958" Nov 25 12:24:27 crc kubenswrapper[4715]: I1125 12:24:27.165913 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312035 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gcfn5"] Nov 25 12:24:31 crc kubenswrapper[4715]: E1125 12:24:31.312659 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerName="extract" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312672 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerName="extract" Nov 25 12:24:31 crc kubenswrapper[4715]: E1125 12:24:31.312690 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerName="extract-utilities" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312697 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerName="extract-utilities" Nov 25 12:24:31 crc kubenswrapper[4715]: E1125 12:24:31.312704 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731e898e-a432-4696-ab69-cc1e0263558e" containerName="extract" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312711 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="731e898e-a432-4696-ab69-cc1e0263558e" containerName="extract" Nov 25 12:24:31 crc kubenswrapper[4715]: E1125 12:24:31.312725 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731e898e-a432-4696-ab69-cc1e0263558e" containerName="util" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312730 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="731e898e-a432-4696-ab69-cc1e0263558e" containerName="util" Nov 25 12:24:31 crc kubenswrapper[4715]: E1125 12:24:31.312744 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerName="util" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312750 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerName="util" Nov 25 12:24:31 crc kubenswrapper[4715]: E1125 12:24:31.312759 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerName="pull" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312764 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerName="pull" Nov 25 12:24:31 crc kubenswrapper[4715]: E1125 12:24:31.312774 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerName="extract-content" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312780 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerName="extract-content" Nov 25 12:24:31 crc kubenswrapper[4715]: E1125 12:24:31.312786 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerName="registry-server" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312792 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerName="registry-server" Nov 25 12:24:31 crc kubenswrapper[4715]: E1125 12:24:31.312802 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731e898e-a432-4696-ab69-cc1e0263558e" containerName="pull" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312807 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="731e898e-a432-4696-ab69-cc1e0263558e" containerName="pull" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312908 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d1aad28-aabe-40fb-8d6c-3d3643a6a52b" containerName="extract" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312917 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="731e898e-a432-4696-ab69-cc1e0263558e" containerName="extract" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.312925 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cea6d8c6-58c6-4abb-b039-af291a90e8be" containerName="registry-server" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.314028 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.321432 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcfn5"] Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.446676 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-utilities\") pod \"redhat-marketplace-gcfn5\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.446867 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnjx4\" (UniqueName: \"kubernetes.io/projected/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-kube-api-access-fnjx4\") pod \"redhat-marketplace-gcfn5\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.447059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-catalog-content\") pod \"redhat-marketplace-gcfn5\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.548611 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-utilities\") pod \"redhat-marketplace-gcfn5\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.548713 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnjx4\" (UniqueName: \"kubernetes.io/projected/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-kube-api-access-fnjx4\") pod \"redhat-marketplace-gcfn5\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.548766 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-catalog-content\") pod \"redhat-marketplace-gcfn5\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.549202 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-utilities\") pod \"redhat-marketplace-gcfn5\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.549305 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-catalog-content\") pod \"redhat-marketplace-gcfn5\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.570488 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnjx4\" (UniqueName: \"kubernetes.io/projected/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-kube-api-access-fnjx4\") pod \"redhat-marketplace-gcfn5\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:31 crc kubenswrapper[4715]: I1125 12:24:31.632086 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:32 crc kubenswrapper[4715]: I1125 12:24:32.045963 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcfn5"] Nov 25 12:24:32 crc kubenswrapper[4715]: I1125 12:24:32.200768 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcfn5" event={"ID":"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd","Type":"ContainerStarted","Data":"fe5b4745689cd8cdb16ed605289bde5373d46107ce99a677b9cfe0e79d99db2b"} Nov 25 12:24:33 crc kubenswrapper[4715]: I1125 12:24:33.207663 4715 generic.go:334] "Generic (PLEG): container finished" podID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerID="1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c" exitCode=0 Nov 25 12:24:33 crc kubenswrapper[4715]: I1125 12:24:33.207794 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcfn5" event={"ID":"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd","Type":"ContainerDied","Data":"1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c"} Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.216080 4715 generic.go:334] "Generic (PLEG): container finished" podID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerID="8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216" exitCode=0 Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.216161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcfn5" event={"ID":"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd","Type":"ContainerDied","Data":"8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216"} Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.423941 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk"] Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.424723 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.428412 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-6rv42" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.429081 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.436686 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk"] Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.585875 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/116af384-b2b3-422d-84a4-38cc7affcf33-webhook-cert\") pod \"horizon-operator-controller-manager-5cf87789bc-srhpk\" (UID: \"116af384-b2b3-422d-84a4-38cc7affcf33\") " pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.585958 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvksq\" (UniqueName: \"kubernetes.io/projected/116af384-b2b3-422d-84a4-38cc7affcf33-kube-api-access-mvksq\") pod \"horizon-operator-controller-manager-5cf87789bc-srhpk\" (UID: \"116af384-b2b3-422d-84a4-38cc7affcf33\") " pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.586045 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/116af384-b2b3-422d-84a4-38cc7affcf33-apiservice-cert\") pod \"horizon-operator-controller-manager-5cf87789bc-srhpk\" (UID: \"116af384-b2b3-422d-84a4-38cc7affcf33\") " pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.686881 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/116af384-b2b3-422d-84a4-38cc7affcf33-webhook-cert\") pod \"horizon-operator-controller-manager-5cf87789bc-srhpk\" (UID: \"116af384-b2b3-422d-84a4-38cc7affcf33\") " pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.686946 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvksq\" (UniqueName: \"kubernetes.io/projected/116af384-b2b3-422d-84a4-38cc7affcf33-kube-api-access-mvksq\") pod \"horizon-operator-controller-manager-5cf87789bc-srhpk\" (UID: \"116af384-b2b3-422d-84a4-38cc7affcf33\") " pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.686989 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/116af384-b2b3-422d-84a4-38cc7affcf33-apiservice-cert\") pod \"horizon-operator-controller-manager-5cf87789bc-srhpk\" (UID: \"116af384-b2b3-422d-84a4-38cc7affcf33\") " pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.692008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/116af384-b2b3-422d-84a4-38cc7affcf33-apiservice-cert\") pod \"horizon-operator-controller-manager-5cf87789bc-srhpk\" (UID: \"116af384-b2b3-422d-84a4-38cc7affcf33\") " pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.692217 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/116af384-b2b3-422d-84a4-38cc7affcf33-webhook-cert\") pod \"horizon-operator-controller-manager-5cf87789bc-srhpk\" (UID: \"116af384-b2b3-422d-84a4-38cc7affcf33\") " pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.710141 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvksq\" (UniqueName: \"kubernetes.io/projected/116af384-b2b3-422d-84a4-38cc7affcf33-kube-api-access-mvksq\") pod \"horizon-operator-controller-manager-5cf87789bc-srhpk\" (UID: \"116af384-b2b3-422d-84a4-38cc7affcf33\") " pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:34 crc kubenswrapper[4715]: I1125 12:24:34.741804 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:35 crc kubenswrapper[4715]: I1125 12:24:35.190348 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk"] Nov 25 12:24:35 crc kubenswrapper[4715]: W1125 12:24:35.197509 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod116af384_b2b3_422d_84a4_38cc7affcf33.slice/crio-81e1d9afe294e5d92a26c8fa19b2c347e10e80aed42a3e03225293bf2d05fb58 WatchSource:0}: Error finding container 81e1d9afe294e5d92a26c8fa19b2c347e10e80aed42a3e03225293bf2d05fb58: Status 404 returned error can't find the container with id 81e1d9afe294e5d92a26c8fa19b2c347e10e80aed42a3e03225293bf2d05fb58 Nov 25 12:24:35 crc kubenswrapper[4715]: I1125 12:24:35.224936 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcfn5" event={"ID":"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd","Type":"ContainerStarted","Data":"46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e"} Nov 25 12:24:35 crc kubenswrapper[4715]: I1125 12:24:35.226278 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" event={"ID":"116af384-b2b3-422d-84a4-38cc7affcf33","Type":"ContainerStarted","Data":"81e1d9afe294e5d92a26c8fa19b2c347e10e80aed42a3e03225293bf2d05fb58"} Nov 25 12:24:35 crc kubenswrapper[4715]: I1125 12:24:35.248278 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gcfn5" podStartSLOduration=2.856418993 podStartE2EDuration="4.248259954s" podCreationTimestamp="2025-11-25 12:24:31 +0000 UTC" firstStartedPulling="2025-11-25 12:24:33.209579915 +0000 UTC m=+923.717082936" lastFinishedPulling="2025-11-25 12:24:34.601420876 +0000 UTC m=+925.108923897" observedRunningTime="2025-11-25 12:24:35.244625909 +0000 UTC m=+925.752128930" watchObservedRunningTime="2025-11-25 12:24:35.248259954 +0000 UTC m=+925.755762975" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.721172 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sl46m"] Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.730842 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.749041 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sl46m"] Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.842890 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-utilities\") pod \"redhat-operators-sl46m\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.842997 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-catalog-content\") pod \"redhat-operators-sl46m\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.843130 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w26d4\" (UniqueName: \"kubernetes.io/projected/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-kube-api-access-w26d4\") pod \"redhat-operators-sl46m\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.944473 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-utilities\") pod \"redhat-operators-sl46m\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.944550 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-catalog-content\") pod \"redhat-operators-sl46m\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.944623 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w26d4\" (UniqueName: \"kubernetes.io/projected/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-kube-api-access-w26d4\") pod \"redhat-operators-sl46m\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.944983 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-utilities\") pod \"redhat-operators-sl46m\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.945256 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-catalog-content\") pod \"redhat-operators-sl46m\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:37 crc kubenswrapper[4715]: I1125 12:24:37.969803 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w26d4\" (UniqueName: \"kubernetes.io/projected/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-kube-api-access-w26d4\") pod \"redhat-operators-sl46m\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:38 crc kubenswrapper[4715]: I1125 12:24:38.056746 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:38 crc kubenswrapper[4715]: I1125 12:24:38.246628 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" event={"ID":"116af384-b2b3-422d-84a4-38cc7affcf33","Type":"ContainerStarted","Data":"cc3991e0b19926bbea7b15b461b2bf51c8c15839f88dfcc57076667bd99f3840"} Nov 25 12:24:38 crc kubenswrapper[4715]: I1125 12:24:38.247225 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:38 crc kubenswrapper[4715]: I1125 12:24:38.274250 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" podStartSLOduration=2.293498862 podStartE2EDuration="4.27423064s" podCreationTimestamp="2025-11-25 12:24:34 +0000 UTC" firstStartedPulling="2025-11-25 12:24:35.199396561 +0000 UTC m=+925.706899582" lastFinishedPulling="2025-11-25 12:24:37.180128329 +0000 UTC m=+927.687631360" observedRunningTime="2025-11-25 12:24:38.271899989 +0000 UTC m=+928.779403020" watchObservedRunningTime="2025-11-25 12:24:38.27423064 +0000 UTC m=+928.781733671" Nov 25 12:24:38 crc kubenswrapper[4715]: I1125 12:24:38.484964 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sl46m"] Nov 25 12:24:38 crc kubenswrapper[4715]: W1125 12:24:38.487094 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50dbe4a5_91ef_43da_a7ca_f2983abdcec4.slice/crio-6f494495a597a9128d1a9061534dd54c7402ebbffba88311c27ce570571a8fa8 WatchSource:0}: Error finding container 6f494495a597a9128d1a9061534dd54c7402ebbffba88311c27ce570571a8fa8: Status 404 returned error can't find the container with id 6f494495a597a9128d1a9061534dd54c7402ebbffba88311c27ce570571a8fa8 Nov 25 12:24:39 crc kubenswrapper[4715]: I1125 12:24:39.256399 4715 generic.go:334] "Generic (PLEG): container finished" podID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerID="9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821" exitCode=0 Nov 25 12:24:39 crc kubenswrapper[4715]: I1125 12:24:39.256457 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sl46m" event={"ID":"50dbe4a5-91ef-43da-a7ca-f2983abdcec4","Type":"ContainerDied","Data":"9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821"} Nov 25 12:24:39 crc kubenswrapper[4715]: I1125 12:24:39.256768 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sl46m" event={"ID":"50dbe4a5-91ef-43da-a7ca-f2983abdcec4","Type":"ContainerStarted","Data":"6f494495a597a9128d1a9061534dd54c7402ebbffba88311c27ce570571a8fa8"} Nov 25 12:24:40 crc kubenswrapper[4715]: I1125 12:24:40.265848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sl46m" event={"ID":"50dbe4a5-91ef-43da-a7ca-f2983abdcec4","Type":"ContainerStarted","Data":"940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae"} Nov 25 12:24:41 crc kubenswrapper[4715]: I1125 12:24:41.015232 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-7f7c6799b4-5dgxc" Nov 25 12:24:41 crc kubenswrapper[4715]: I1125 12:24:41.272160 4715 generic.go:334] "Generic (PLEG): container finished" podID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerID="940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae" exitCode=0 Nov 25 12:24:41 crc kubenswrapper[4715]: I1125 12:24:41.272249 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sl46m" event={"ID":"50dbe4a5-91ef-43da-a7ca-f2983abdcec4","Type":"ContainerDied","Data":"940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae"} Nov 25 12:24:41 crc kubenswrapper[4715]: I1125 12:24:41.274126 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 12:24:41 crc kubenswrapper[4715]: I1125 12:24:41.633125 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:41 crc kubenswrapper[4715]: I1125 12:24:41.633179 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:41 crc kubenswrapper[4715]: I1125 12:24:41.672988 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:42 crc kubenswrapper[4715]: I1125 12:24:42.280589 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sl46m" event={"ID":"50dbe4a5-91ef-43da-a7ca-f2983abdcec4","Type":"ContainerStarted","Data":"e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24"} Nov 25 12:24:42 crc kubenswrapper[4715]: I1125 12:24:42.299201 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sl46m" podStartSLOduration=2.899153935 podStartE2EDuration="5.299166352s" podCreationTimestamp="2025-11-25 12:24:37 +0000 UTC" firstStartedPulling="2025-11-25 12:24:39.257908214 +0000 UTC m=+929.765411235" lastFinishedPulling="2025-11-25 12:24:41.657920631 +0000 UTC m=+932.165423652" observedRunningTime="2025-11-25 12:24:42.297453337 +0000 UTC m=+932.804956358" watchObservedRunningTime="2025-11-25 12:24:42.299166352 +0000 UTC m=+932.806669373" Nov 25 12:24:42 crc kubenswrapper[4715]: I1125 12:24:42.326034 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:43 crc kubenswrapper[4715]: I1125 12:24:43.511883 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcfn5"] Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.292525 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gcfn5" podUID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerName="registry-server" containerID="cri-o://46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e" gracePeriod=2 Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.590687 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58"] Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.591587 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.595763 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.595970 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-m9m4x" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.596035 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58"] Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.735854 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.735913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/436fab16-f756-4564-ad51-316c4e869463-apiservice-cert\") pod \"swift-operator-controller-manager-7d7fdbdb55-hsz58\" (UID: \"436fab16-f756-4564-ad51-316c4e869463\") " pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.735981 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/436fab16-f756-4564-ad51-316c4e869463-webhook-cert\") pod \"swift-operator-controller-manager-7d7fdbdb55-hsz58\" (UID: \"436fab16-f756-4564-ad51-316c4e869463\") " pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.736017 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkmtf\" (UniqueName: \"kubernetes.io/projected/436fab16-f756-4564-ad51-316c4e869463-kube-api-access-xkmtf\") pod \"swift-operator-controller-manager-7d7fdbdb55-hsz58\" (UID: \"436fab16-f756-4564-ad51-316c4e869463\") " pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.748340 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5cf87789bc-srhpk" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.837139 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-utilities\") pod \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.837231 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-catalog-content\") pod \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.837340 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnjx4\" (UniqueName: \"kubernetes.io/projected/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-kube-api-access-fnjx4\") pod \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\" (UID: \"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd\") " Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.837621 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/436fab16-f756-4564-ad51-316c4e869463-webhook-cert\") pod \"swift-operator-controller-manager-7d7fdbdb55-hsz58\" (UID: \"436fab16-f756-4564-ad51-316c4e869463\") " pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.837672 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkmtf\" (UniqueName: \"kubernetes.io/projected/436fab16-f756-4564-ad51-316c4e869463-kube-api-access-xkmtf\") pod \"swift-operator-controller-manager-7d7fdbdb55-hsz58\" (UID: \"436fab16-f756-4564-ad51-316c4e869463\") " pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.837796 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/436fab16-f756-4564-ad51-316c4e869463-apiservice-cert\") pod \"swift-operator-controller-manager-7d7fdbdb55-hsz58\" (UID: \"436fab16-f756-4564-ad51-316c4e869463\") " pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.837910 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-utilities" (OuterVolumeSpecName: "utilities") pod "73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" (UID: "73c8ce8c-c8c4-4600-a806-c3d1ae1330dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.843658 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/436fab16-f756-4564-ad51-316c4e869463-apiservice-cert\") pod \"swift-operator-controller-manager-7d7fdbdb55-hsz58\" (UID: \"436fab16-f756-4564-ad51-316c4e869463\") " pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.843731 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/436fab16-f756-4564-ad51-316c4e869463-webhook-cert\") pod \"swift-operator-controller-manager-7d7fdbdb55-hsz58\" (UID: \"436fab16-f756-4564-ad51-316c4e869463\") " pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.846312 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-kube-api-access-fnjx4" (OuterVolumeSpecName: "kube-api-access-fnjx4") pod "73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" (UID: "73c8ce8c-c8c4-4600-a806-c3d1ae1330dd"). InnerVolumeSpecName "kube-api-access-fnjx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.853974 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkmtf\" (UniqueName: \"kubernetes.io/projected/436fab16-f756-4564-ad51-316c4e869463-kube-api-access-xkmtf\") pod \"swift-operator-controller-manager-7d7fdbdb55-hsz58\" (UID: \"436fab16-f756-4564-ad51-316c4e869463\") " pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.854257 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" (UID: "73c8ce8c-c8c4-4600-a806-c3d1ae1330dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.921351 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.939650 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnjx4\" (UniqueName: \"kubernetes.io/projected/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-kube-api-access-fnjx4\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.939692 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:44 crc kubenswrapper[4715]: I1125 12:24:44.939704 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.302342 4715 generic.go:334] "Generic (PLEG): container finished" podID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerID="46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e" exitCode=0 Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.302447 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcfn5" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.302470 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcfn5" event={"ID":"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd","Type":"ContainerDied","Data":"46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e"} Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.302646 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcfn5" event={"ID":"73c8ce8c-c8c4-4600-a806-c3d1ae1330dd","Type":"ContainerDied","Data":"fe5b4745689cd8cdb16ed605289bde5373d46107ce99a677b9cfe0e79d99db2b"} Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.302695 4715 scope.go:117] "RemoveContainer" containerID="46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.311652 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58"] Nov 25 12:24:45 crc kubenswrapper[4715]: W1125 12:24:45.317248 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod436fab16_f756_4564_ad51_316c4e869463.slice/crio-2ed7e558f5d0586e13c1df7fa543a3c886cab19e7b29e9b4c7d7ff540f8630b0 WatchSource:0}: Error finding container 2ed7e558f5d0586e13c1df7fa543a3c886cab19e7b29e9b4c7d7ff540f8630b0: Status 404 returned error can't find the container with id 2ed7e558f5d0586e13c1df7fa543a3c886cab19e7b29e9b4c7d7ff540f8630b0 Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.324561 4715 scope.go:117] "RemoveContainer" containerID="8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.336618 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcfn5"] Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.343001 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcfn5"] Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.358108 4715 scope.go:117] "RemoveContainer" containerID="1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.374746 4715 scope.go:117] "RemoveContainer" containerID="46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e" Nov 25 12:24:45 crc kubenswrapper[4715]: E1125 12:24:45.375242 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e\": container with ID starting with 46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e not found: ID does not exist" containerID="46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.375288 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e"} err="failed to get container status \"46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e\": rpc error: code = NotFound desc = could not find container \"46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e\": container with ID starting with 46b6a0bd0be5d53ad33dbf5b45e8fa5f310f3271715d68f3637aa5f509a1e96e not found: ID does not exist" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.375317 4715 scope.go:117] "RemoveContainer" containerID="8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216" Nov 25 12:24:45 crc kubenswrapper[4715]: E1125 12:24:45.375694 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216\": container with ID starting with 8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216 not found: ID does not exist" containerID="8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.375723 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216"} err="failed to get container status \"8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216\": rpc error: code = NotFound desc = could not find container \"8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216\": container with ID starting with 8674674569daf758f61f720b9329a8fc132f0d0062082d108b2541228bf70216 not found: ID does not exist" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.375745 4715 scope.go:117] "RemoveContainer" containerID="1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c" Nov 25 12:24:45 crc kubenswrapper[4715]: E1125 12:24:45.376103 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c\": container with ID starting with 1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c not found: ID does not exist" containerID="1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c" Nov 25 12:24:45 crc kubenswrapper[4715]: I1125 12:24:45.376158 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c"} err="failed to get container status \"1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c\": rpc error: code = NotFound desc = could not find container \"1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c\": container with ID starting with 1cd2324b356748673fd50beb20230b30fa1f9a387a11b59594e55d6d8b78362c not found: ID does not exist" Nov 25 12:24:46 crc kubenswrapper[4715]: I1125 12:24:46.320862 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" event={"ID":"436fab16-f756-4564-ad51-316c4e869463","Type":"ContainerStarted","Data":"2ed7e558f5d0586e13c1df7fa543a3c886cab19e7b29e9b4c7d7ff540f8630b0"} Nov 25 12:24:46 crc kubenswrapper[4715]: I1125 12:24:46.701555 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" path="/var/lib/kubelet/pods/73c8ce8c-c8c4-4600-a806-c3d1ae1330dd/volumes" Nov 25 12:24:48 crc kubenswrapper[4715]: I1125 12:24:48.057867 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:48 crc kubenswrapper[4715]: I1125 12:24:48.058247 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:48 crc kubenswrapper[4715]: I1125 12:24:48.172852 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:48 crc kubenswrapper[4715]: I1125 12:24:48.339963 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" event={"ID":"436fab16-f756-4564-ad51-316c4e869463","Type":"ContainerStarted","Data":"72e88a57f5074f7fb1efa066617155a444e41d74afb606be8bd4c35410cf4f7f"} Nov 25 12:24:48 crc kubenswrapper[4715]: I1125 12:24:48.356120 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" podStartSLOduration=2.588444539 podStartE2EDuration="4.356104888s" podCreationTimestamp="2025-11-25 12:24:44 +0000 UTC" firstStartedPulling="2025-11-25 12:24:45.324702488 +0000 UTC m=+935.832205509" lastFinishedPulling="2025-11-25 12:24:47.092362837 +0000 UTC m=+937.599865858" observedRunningTime="2025-11-25 12:24:48.353877929 +0000 UTC m=+938.861380950" watchObservedRunningTime="2025-11-25 12:24:48.356104888 +0000 UTC m=+938.863607909" Nov 25 12:24:48 crc kubenswrapper[4715]: I1125 12:24:48.379568 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:49 crc kubenswrapper[4715]: I1125 12:24:49.345606 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.304265 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sl46m"] Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.304771 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sl46m" podUID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerName="registry-server" containerID="cri-o://e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24" gracePeriod=2 Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.680907 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.747530 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-catalog-content\") pod \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.747650 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w26d4\" (UniqueName: \"kubernetes.io/projected/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-kube-api-access-w26d4\") pod \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.747776 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-utilities\") pod \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\" (UID: \"50dbe4a5-91ef-43da-a7ca-f2983abdcec4\") " Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.749032 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-utilities" (OuterVolumeSpecName: "utilities") pod "50dbe4a5-91ef-43da-a7ca-f2983abdcec4" (UID: "50dbe4a5-91ef-43da-a7ca-f2983abdcec4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.753669 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-kube-api-access-w26d4" (OuterVolumeSpecName: "kube-api-access-w26d4") pod "50dbe4a5-91ef-43da-a7ca-f2983abdcec4" (UID: "50dbe4a5-91ef-43da-a7ca-f2983abdcec4"). InnerVolumeSpecName "kube-api-access-w26d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.835134 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50dbe4a5-91ef-43da-a7ca-f2983abdcec4" (UID: "50dbe4a5-91ef-43da-a7ca-f2983abdcec4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.849198 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.849259 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w26d4\" (UniqueName: \"kubernetes.io/projected/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-kube-api-access-w26d4\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:52 crc kubenswrapper[4715]: I1125 12:24:52.849276 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50dbe4a5-91ef-43da-a7ca-f2983abdcec4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.376849 4715 generic.go:334] "Generic (PLEG): container finished" podID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerID="e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24" exitCode=0 Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.377022 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sl46m" event={"ID":"50dbe4a5-91ef-43da-a7ca-f2983abdcec4","Type":"ContainerDied","Data":"e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24"} Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.377428 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sl46m" event={"ID":"50dbe4a5-91ef-43da-a7ca-f2983abdcec4","Type":"ContainerDied","Data":"6f494495a597a9128d1a9061534dd54c7402ebbffba88311c27ce570571a8fa8"} Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.377464 4715 scope.go:117] "RemoveContainer" containerID="e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.377207 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sl46m" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.406826 4715 scope.go:117] "RemoveContainer" containerID="940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.420531 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sl46m"] Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.427346 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sl46m"] Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.435064 4715 scope.go:117] "RemoveContainer" containerID="9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.456967 4715 scope.go:117] "RemoveContainer" containerID="e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24" Nov 25 12:24:53 crc kubenswrapper[4715]: E1125 12:24:53.457525 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24\": container with ID starting with e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24 not found: ID does not exist" containerID="e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.457568 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24"} err="failed to get container status \"e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24\": rpc error: code = NotFound desc = could not find container \"e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24\": container with ID starting with e709cda16574279e39d4d27c16fe965b2f1968fa34cf67ec3a7fc2a081200a24 not found: ID does not exist" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.457590 4715 scope.go:117] "RemoveContainer" containerID="940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae" Nov 25 12:24:53 crc kubenswrapper[4715]: E1125 12:24:53.457884 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae\": container with ID starting with 940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae not found: ID does not exist" containerID="940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.457905 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae"} err="failed to get container status \"940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae\": rpc error: code = NotFound desc = could not find container \"940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae\": container with ID starting with 940176823dcef620d52d1448b67c05d8041e3ab9f71f726c33acdb032c1d44ae not found: ID does not exist" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.457920 4715 scope.go:117] "RemoveContainer" containerID="9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821" Nov 25 12:24:53 crc kubenswrapper[4715]: E1125 12:24:53.458472 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821\": container with ID starting with 9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821 not found: ID does not exist" containerID="9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821" Nov 25 12:24:53 crc kubenswrapper[4715]: I1125 12:24:53.458512 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821"} err="failed to get container status \"9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821\": rpc error: code = NotFound desc = could not find container \"9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821\": container with ID starting with 9fce8a51d32baa8ab4923c2177c8a1a673cb7a559c83520c10836144f1456821 not found: ID does not exist" Nov 25 12:24:54 crc kubenswrapper[4715]: I1125 12:24:54.704068 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" path="/var/lib/kubelet/pods/50dbe4a5-91ef-43da-a7ca-f2983abdcec4/volumes" Nov 25 12:24:54 crc kubenswrapper[4715]: I1125 12:24:54.925625 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7d7fdbdb55-hsz58" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.647103 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 12:24:59 crc kubenswrapper[4715]: E1125 12:24:59.647894 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerName="registry-server" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.647909 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerName="registry-server" Nov 25 12:24:59 crc kubenswrapper[4715]: E1125 12:24:59.647922 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerName="extract-content" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.647929 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerName="extract-content" Nov 25 12:24:59 crc kubenswrapper[4715]: E1125 12:24:59.647939 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerName="registry-server" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.647947 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerName="registry-server" Nov 25 12:24:59 crc kubenswrapper[4715]: E1125 12:24:59.647959 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerName="extract-utilities" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.647968 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerName="extract-utilities" Nov 25 12:24:59 crc kubenswrapper[4715]: E1125 12:24:59.647980 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerName="extract-content" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.647986 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerName="extract-content" Nov 25 12:24:59 crc kubenswrapper[4715]: E1125 12:24:59.648002 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerName="extract-utilities" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.648009 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerName="extract-utilities" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.648200 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="73c8ce8c-c8c4-4600-a806-c3d1ae1330dd" containerName="registry-server" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.648220 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="50dbe4a5-91ef-43da-a7ca-f2983abdcec4" containerName="registry-server" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.652165 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.654611 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.654612 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.655054 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-z6k77" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.655054 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.676408 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.746302 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.746349 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.746369 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5jl7\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-kube-api-access-j5jl7\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.746407 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edd3d91a-8a99-4a84-880f-75c3968b264e-lock\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.746439 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edd3d91a-8a99-4a84-880f-75c3968b264e-cache\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.849149 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.849230 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.849261 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5jl7\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-kube-api-access-j5jl7\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.849312 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edd3d91a-8a99-4a84-880f-75c3968b264e-lock\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.849356 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edd3d91a-8a99-4a84-880f-75c3968b264e-cache\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: E1125 12:24:59.849415 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:24:59 crc kubenswrapper[4715]: E1125 12:24:59.849434 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.849430 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: E1125 12:24:59.849484 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift podName:edd3d91a-8a99-4a84-880f-75c3968b264e nodeName:}" failed. No retries permitted until 2025-11-25 12:25:00.349465291 +0000 UTC m=+950.856968312 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift") pod "swift-storage-0" (UID: "edd3d91a-8a99-4a84-880f-75c3968b264e") : configmap "swift-ring-files" not found Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.849902 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edd3d91a-8a99-4a84-880f-75c3968b264e-cache\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.850436 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edd3d91a-8a99-4a84-880f-75c3968b264e-lock\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.870531 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:24:59 crc kubenswrapper[4715]: I1125 12:24:59.877773 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5jl7\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-kube-api-access-j5jl7\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.355844 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:25:00 crc kubenswrapper[4715]: E1125 12:25:00.356074 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:25:00 crc kubenswrapper[4715]: E1125 12:25:00.356288 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 12:25:00 crc kubenswrapper[4715]: E1125 12:25:00.356411 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift podName:edd3d91a-8a99-4a84-880f-75c3968b264e nodeName:}" failed. No retries permitted until 2025-11-25 12:25:01.356387839 +0000 UTC m=+951.863890870 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift") pod "swift-storage-0" (UID: "edd3d91a-8a99-4a84-880f-75c3968b264e") : configmap "swift-ring-files" not found Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.557564 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n"] Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.559054 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.560778 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.573973 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n"] Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.660232 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552e79b9-1232-48cb-8f68-02488b8d34e7-run-httpd\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.660339 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552e79b9-1232-48cb-8f68-02488b8d34e7-log-httpd\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.660363 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552e79b9-1232-48cb-8f68-02488b8d34e7-config-data\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.660387 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gqkw\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-kube-api-access-4gqkw\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.660468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.762075 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552e79b9-1232-48cb-8f68-02488b8d34e7-log-httpd\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.762122 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552e79b9-1232-48cb-8f68-02488b8d34e7-config-data\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.762143 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gqkw\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-kube-api-access-4gqkw\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.762210 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.762308 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552e79b9-1232-48cb-8f68-02488b8d34e7-run-httpd\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.762645 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552e79b9-1232-48cb-8f68-02488b8d34e7-log-httpd\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: E1125 12:25:00.763384 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:25:00 crc kubenswrapper[4715]: E1125 12:25:00.763408 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n: configmap "swift-ring-files" not found Nov 25 12:25:00 crc kubenswrapper[4715]: E1125 12:25:00.763448 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift podName:552e79b9-1232-48cb-8f68-02488b8d34e7 nodeName:}" failed. No retries permitted until 2025-11-25 12:25:01.263433246 +0000 UTC m=+951.770936267 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift") pod "swift-proxy-6bd58cfcf7-jlm2n" (UID: "552e79b9-1232-48cb-8f68-02488b8d34e7") : configmap "swift-ring-files" not found Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.763854 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552e79b9-1232-48cb-8f68-02488b8d34e7-run-httpd\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.767857 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552e79b9-1232-48cb-8f68-02488b8d34e7-config-data\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:00 crc kubenswrapper[4715]: I1125 12:25:00.780618 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gqkw\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-kube-api-access-4gqkw\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.277933 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:01 crc kubenswrapper[4715]: E1125 12:25:01.278178 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:25:01 crc kubenswrapper[4715]: E1125 12:25:01.278494 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n: configmap "swift-ring-files" not found Nov 25 12:25:01 crc kubenswrapper[4715]: E1125 12:25:01.278554 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift podName:552e79b9-1232-48cb-8f68-02488b8d34e7 nodeName:}" failed. No retries permitted until 2025-11-25 12:25:02.278537298 +0000 UTC m=+952.786040329 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift") pod "swift-proxy-6bd58cfcf7-jlm2n" (UID: "552e79b9-1232-48cb-8f68-02488b8d34e7") : configmap "swift-ring-files" not found Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.318344 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-fkgk9"] Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.319267 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-fkgk9" Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.321600 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-x6bhj" Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.338686 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-fkgk9"] Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.379998 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:25:01 crc kubenswrapper[4715]: E1125 12:25:01.380204 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:25:01 crc kubenswrapper[4715]: E1125 12:25:01.380222 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 12:25:01 crc kubenswrapper[4715]: E1125 12:25:01.380273 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift podName:edd3d91a-8a99-4a84-880f-75c3968b264e nodeName:}" failed. No retries permitted until 2025-11-25 12:25:03.380255526 +0000 UTC m=+953.887758547 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift") pod "swift-storage-0" (UID: "edd3d91a-8a99-4a84-880f-75c3968b264e") : configmap "swift-ring-files" not found Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.481284 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqpr4\" (UniqueName: \"kubernetes.io/projected/6de0e4e1-4bcd-4115-9c83-0827f880f400-kube-api-access-kqpr4\") pod \"glance-operator-index-fkgk9\" (UID: \"6de0e4e1-4bcd-4115-9c83-0827f880f400\") " pod="openstack-operators/glance-operator-index-fkgk9" Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.582681 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqpr4\" (UniqueName: \"kubernetes.io/projected/6de0e4e1-4bcd-4115-9c83-0827f880f400-kube-api-access-kqpr4\") pod \"glance-operator-index-fkgk9\" (UID: \"6de0e4e1-4bcd-4115-9c83-0827f880f400\") " pod="openstack-operators/glance-operator-index-fkgk9" Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.604273 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqpr4\" (UniqueName: \"kubernetes.io/projected/6de0e4e1-4bcd-4115-9c83-0827f880f400-kube-api-access-kqpr4\") pod \"glance-operator-index-fkgk9\" (UID: \"6de0e4e1-4bcd-4115-9c83-0827f880f400\") " pod="openstack-operators/glance-operator-index-fkgk9" Nov 25 12:25:01 crc kubenswrapper[4715]: I1125 12:25:01.638211 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-fkgk9" Nov 25 12:25:02 crc kubenswrapper[4715]: I1125 12:25:02.050345 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-fkgk9"] Nov 25 12:25:02 crc kubenswrapper[4715]: I1125 12:25:02.293605 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:02 crc kubenswrapper[4715]: E1125 12:25:02.293794 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:25:02 crc kubenswrapper[4715]: E1125 12:25:02.293813 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n: configmap "swift-ring-files" not found Nov 25 12:25:02 crc kubenswrapper[4715]: E1125 12:25:02.293867 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift podName:552e79b9-1232-48cb-8f68-02488b8d34e7 nodeName:}" failed. No retries permitted until 2025-11-25 12:25:04.293849112 +0000 UTC m=+954.801352133 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift") pod "swift-proxy-6bd58cfcf7-jlm2n" (UID: "552e79b9-1232-48cb-8f68-02488b8d34e7") : configmap "swift-ring-files" not found Nov 25 12:25:02 crc kubenswrapper[4715]: I1125 12:25:02.435507 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-fkgk9" event={"ID":"6de0e4e1-4bcd-4115-9c83-0827f880f400","Type":"ContainerStarted","Data":"53e3a9e05a25091b289041f317aae398c9c02181cfbf9a15978544294bc0c8a8"} Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.408550 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:25:03 crc kubenswrapper[4715]: E1125 12:25:03.408729 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:25:03 crc kubenswrapper[4715]: E1125 12:25:03.408941 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 12:25:03 crc kubenswrapper[4715]: E1125 12:25:03.409012 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift podName:edd3d91a-8a99-4a84-880f-75c3968b264e nodeName:}" failed. No retries permitted until 2025-11-25 12:25:07.408989564 +0000 UTC m=+957.916492585 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift") pod "swift-storage-0" (UID: "edd3d91a-8a99-4a84-880f-75c3968b264e") : configmap "swift-ring-files" not found Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.649007 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-ll67x"] Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.649972 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.651938 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.653323 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.659328 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-ll67x"] Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.712764 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-ring-data-devices\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.712897 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9070852d-f5ea-415a-b4f5-1c0537c628f8-etc-swift\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.712942 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dmhr\" (UniqueName: \"kubernetes.io/projected/9070852d-f5ea-415a-b4f5-1c0537c628f8-kube-api-access-4dmhr\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.712974 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-scripts\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.713276 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-dispersionconf\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.713333 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-swiftconf\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.814623 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9070852d-f5ea-415a-b4f5-1c0537c628f8-etc-swift\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.814674 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dmhr\" (UniqueName: \"kubernetes.io/projected/9070852d-f5ea-415a-b4f5-1c0537c628f8-kube-api-access-4dmhr\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.814699 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-scripts\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.814753 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-swiftconf\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.814772 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-dispersionconf\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.814843 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-ring-data-devices\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.815618 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-ring-data-devices\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.815850 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9070852d-f5ea-415a-b4f5-1c0537c628f8-etc-swift\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.818103 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-scripts\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.823151 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-dispersionconf\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.823201 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-swiftconf\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.832111 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dmhr\" (UniqueName: \"kubernetes.io/projected/9070852d-f5ea-415a-b4f5-1c0537c628f8-kube-api-access-4dmhr\") pod \"swift-ring-rebalance-ll67x\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:03 crc kubenswrapper[4715]: I1125 12:25:03.966335 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:04 crc kubenswrapper[4715]: I1125 12:25:04.321177 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:04 crc kubenswrapper[4715]: E1125 12:25:04.321403 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:25:04 crc kubenswrapper[4715]: E1125 12:25:04.321599 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n: configmap "swift-ring-files" not found Nov 25 12:25:04 crc kubenswrapper[4715]: E1125 12:25:04.321651 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift podName:552e79b9-1232-48cb-8f68-02488b8d34e7 nodeName:}" failed. No retries permitted until 2025-11-25 12:25:08.321636045 +0000 UTC m=+958.829139066 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift") pod "swift-proxy-6bd58cfcf7-jlm2n" (UID: "552e79b9-1232-48cb-8f68-02488b8d34e7") : configmap "swift-ring-files" not found Nov 25 12:25:04 crc kubenswrapper[4715]: I1125 12:25:04.454639 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-fkgk9" event={"ID":"6de0e4e1-4bcd-4115-9c83-0827f880f400","Type":"ContainerStarted","Data":"69d91bbce7e33d2a57500ef6c04bf81f79a4a1d1e78585c27c94d99ba8debb43"} Nov 25 12:25:04 crc kubenswrapper[4715]: W1125 12:25:04.463556 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9070852d_f5ea_415a_b4f5_1c0537c628f8.slice/crio-a082d464a73943704df7a6d36fae40475949ef7a22382db985c03891e3cbc743 WatchSource:0}: Error finding container a082d464a73943704df7a6d36fae40475949ef7a22382db985c03891e3cbc743: Status 404 returned error can't find the container with id a082d464a73943704df7a6d36fae40475949ef7a22382db985c03891e3cbc743 Nov 25 12:25:04 crc kubenswrapper[4715]: I1125 12:25:04.470180 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-ll67x"] Nov 25 12:25:04 crc kubenswrapper[4715]: I1125 12:25:04.478638 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-fkgk9" podStartSLOduration=1.4601105140000001 podStartE2EDuration="3.478620523s" podCreationTimestamp="2025-11-25 12:25:01 +0000 UTC" firstStartedPulling="2025-11-25 12:25:02.056727482 +0000 UTC m=+952.564230503" lastFinishedPulling="2025-11-25 12:25:04.075237491 +0000 UTC m=+954.582740512" observedRunningTime="2025-11-25 12:25:04.47623462 +0000 UTC m=+954.983737641" watchObservedRunningTime="2025-11-25 12:25:04.478620523 +0000 UTC m=+954.986123544" Nov 25 12:25:05 crc kubenswrapper[4715]: I1125 12:25:05.461947 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" event={"ID":"9070852d-f5ea-415a-b4f5-1c0537c628f8","Type":"ContainerStarted","Data":"a082d464a73943704df7a6d36fae40475949ef7a22382db985c03891e3cbc743"} Nov 25 12:25:07 crc kubenswrapper[4715]: I1125 12:25:07.469155 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:25:07 crc kubenswrapper[4715]: E1125 12:25:07.469519 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:25:07 crc kubenswrapper[4715]: E1125 12:25:07.469548 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 25 12:25:07 crc kubenswrapper[4715]: E1125 12:25:07.469609 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift podName:edd3d91a-8a99-4a84-880f-75c3968b264e nodeName:}" failed. No retries permitted until 2025-11-25 12:25:15.469586952 +0000 UTC m=+965.977089973 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift") pod "swift-storage-0" (UID: "edd3d91a-8a99-4a84-880f-75c3968b264e") : configmap "swift-ring-files" not found Nov 25 12:25:08 crc kubenswrapper[4715]: I1125 12:25:08.384425 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:08 crc kubenswrapper[4715]: E1125 12:25:08.384665 4715 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 25 12:25:08 crc kubenswrapper[4715]: E1125 12:25:08.384809 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n: configmap "swift-ring-files" not found Nov 25 12:25:08 crc kubenswrapper[4715]: E1125 12:25:08.384862 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift podName:552e79b9-1232-48cb-8f68-02488b8d34e7 nodeName:}" failed. No retries permitted until 2025-11-25 12:25:16.384846511 +0000 UTC m=+966.892349532 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift") pod "swift-proxy-6bd58cfcf7-jlm2n" (UID: "552e79b9-1232-48cb-8f68-02488b8d34e7") : configmap "swift-ring-files" not found Nov 25 12:25:08 crc kubenswrapper[4715]: I1125 12:25:08.496287 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" event={"ID":"9070852d-f5ea-415a-b4f5-1c0537c628f8","Type":"ContainerStarted","Data":"5a9f66df52eb3db59a55d33ff29fb154a36b5160ded1dd0a84343d00c4cf77db"} Nov 25 12:25:08 crc kubenswrapper[4715]: I1125 12:25:08.517310 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" podStartSLOduration=2.439011246 podStartE2EDuration="5.517294495s" podCreationTimestamp="2025-11-25 12:25:03 +0000 UTC" firstStartedPulling="2025-11-25 12:25:04.465817627 +0000 UTC m=+954.973320648" lastFinishedPulling="2025-11-25 12:25:07.544100876 +0000 UTC m=+958.051603897" observedRunningTime="2025-11-25 12:25:08.51099136 +0000 UTC m=+959.018494371" watchObservedRunningTime="2025-11-25 12:25:08.517294495 +0000 UTC m=+959.024797516" Nov 25 12:25:11 crc kubenswrapper[4715]: I1125 12:25:11.639496 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-fkgk9" Nov 25 12:25:11 crc kubenswrapper[4715]: I1125 12:25:11.639898 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-fkgk9" Nov 25 12:25:11 crc kubenswrapper[4715]: I1125 12:25:11.671265 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-fkgk9" Nov 25 12:25:12 crc kubenswrapper[4715]: I1125 12:25:12.550841 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-fkgk9" Nov 25 12:25:14 crc kubenswrapper[4715]: I1125 12:25:14.535908 4715 generic.go:334] "Generic (PLEG): container finished" podID="9070852d-f5ea-415a-b4f5-1c0537c628f8" containerID="5a9f66df52eb3db59a55d33ff29fb154a36b5160ded1dd0a84343d00c4cf77db" exitCode=0 Nov 25 12:25:14 crc kubenswrapper[4715]: I1125 12:25:14.535992 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" event={"ID":"9070852d-f5ea-415a-b4f5-1c0537c628f8","Type":"ContainerDied","Data":"5a9f66df52eb3db59a55d33ff29fb154a36b5160ded1dd0a84343d00c4cf77db"} Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.494139 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.505179 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edd3d91a-8a99-4a84-880f-75c3968b264e-etc-swift\") pod \"swift-storage-0\" (UID: \"edd3d91a-8a99-4a84-880f-75c3968b264e\") " pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.568604 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.914524 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.953326 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.972214 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs"] Nov 25 12:25:15 crc kubenswrapper[4715]: E1125 12:25:15.972485 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9070852d-f5ea-415a-b4f5-1c0537c628f8" containerName="swift-ring-rebalance" Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.972497 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9070852d-f5ea-415a-b4f5-1c0537c628f8" containerName="swift-ring-rebalance" Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.972614 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9070852d-f5ea-415a-b4f5-1c0537c628f8" containerName="swift-ring-rebalance" Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.973454 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.977105 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7xsfb" Nov 25 12:25:15 crc kubenswrapper[4715]: I1125 12:25:15.990460 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs"] Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.001411 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dmhr\" (UniqueName: \"kubernetes.io/projected/9070852d-f5ea-415a-b4f5-1c0537c628f8-kube-api-access-4dmhr\") pod \"9070852d-f5ea-415a-b4f5-1c0537c628f8\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.001467 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-ring-data-devices\") pod \"9070852d-f5ea-415a-b4f5-1c0537c628f8\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.001517 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-dispersionconf\") pod \"9070852d-f5ea-415a-b4f5-1c0537c628f8\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.001572 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-swiftconf\") pod \"9070852d-f5ea-415a-b4f5-1c0537c628f8\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.001618 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-scripts\") pod \"9070852d-f5ea-415a-b4f5-1c0537c628f8\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.001659 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9070852d-f5ea-415a-b4f5-1c0537c628f8-etc-swift\") pod \"9070852d-f5ea-415a-b4f5-1c0537c628f8\" (UID: \"9070852d-f5ea-415a-b4f5-1c0537c628f8\") " Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.002514 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9070852d-f5ea-415a-b4f5-1c0537c628f8" (UID: "9070852d-f5ea-415a-b4f5-1c0537c628f8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.003205 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9070852d-f5ea-415a-b4f5-1c0537c628f8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9070852d-f5ea-415a-b4f5-1c0537c628f8" (UID: "9070852d-f5ea-415a-b4f5-1c0537c628f8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.006368 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9070852d-f5ea-415a-b4f5-1c0537c628f8-kube-api-access-4dmhr" (OuterVolumeSpecName: "kube-api-access-4dmhr") pod "9070852d-f5ea-415a-b4f5-1c0537c628f8" (UID: "9070852d-f5ea-415a-b4f5-1c0537c628f8"). InnerVolumeSpecName "kube-api-access-4dmhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.019793 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9070852d-f5ea-415a-b4f5-1c0537c628f8" (UID: "9070852d-f5ea-415a-b4f5-1c0537c628f8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.022161 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-scripts" (OuterVolumeSpecName: "scripts") pod "9070852d-f5ea-415a-b4f5-1c0537c628f8" (UID: "9070852d-f5ea-415a-b4f5-1c0537c628f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.025514 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9070852d-f5ea-415a-b4f5-1c0537c628f8" (UID: "9070852d-f5ea-415a-b4f5-1c0537c628f8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.103035 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-bundle\") pod \"ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.103378 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-util\") pod \"ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.103545 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbcpr\" (UniqueName: \"kubernetes.io/projected/8ff1994e-54dc-4697-9b2f-402bfb693383-kube-api-access-xbcpr\") pod \"ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.104440 4715 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.104491 4715 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9070852d-f5ea-415a-b4f5-1c0537c628f8-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.104504 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.104516 4715 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9070852d-f5ea-415a-b4f5-1c0537c628f8-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.104987 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dmhr\" (UniqueName: \"kubernetes.io/projected/9070852d-f5ea-415a-b4f5-1c0537c628f8-kube-api-access-4dmhr\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.105019 4715 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9070852d-f5ea-415a-b4f5-1c0537c628f8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.206648 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbcpr\" (UniqueName: \"kubernetes.io/projected/8ff1994e-54dc-4697-9b2f-402bfb693383-kube-api-access-xbcpr\") pod \"ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.206762 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-bundle\") pod \"ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.206782 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-util\") pod \"ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.207258 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-util\") pod \"ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.207410 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-bundle\") pod \"ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.227481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbcpr\" (UniqueName: \"kubernetes.io/projected/8ff1994e-54dc-4697-9b2f-402bfb693383-kube-api-access-xbcpr\") pod \"ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.293943 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.410039 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.414530 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/552e79b9-1232-48cb-8f68-02488b8d34e7-etc-swift\") pod \"swift-proxy-6bd58cfcf7-jlm2n\" (UID: \"552e79b9-1232-48cb-8f68-02488b8d34e7\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.480097 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.556022 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.556841 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-ll67x" event={"ID":"9070852d-f5ea-415a-b4f5-1c0537c628f8","Type":"ContainerDied","Data":"a082d464a73943704df7a6d36fae40475949ef7a22382db985c03891e3cbc743"} Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.556915 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a082d464a73943704df7a6d36fae40475949ef7a22382db985c03891e3cbc743" Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.561634 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"ff2f08f9146ed8b5ad25c07fa6d8c02f5726db5afe50a14ce2564dbb45764d90"} Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.720398 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs"] Nov 25 12:25:16 crc kubenswrapper[4715]: I1125 12:25:16.897317 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n"] Nov 25 12:25:16 crc kubenswrapper[4715]: W1125 12:25:16.907531 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod552e79b9_1232_48cb_8f68_02488b8d34e7.slice/crio-9903ea44e7f5a645d518290efdc04e4a2401f1085af8d9c12a8bda27375d1f7e WatchSource:0}: Error finding container 9903ea44e7f5a645d518290efdc04e4a2401f1085af8d9c12a8bda27375d1f7e: Status 404 returned error can't find the container with id 9903ea44e7f5a645d518290efdc04e4a2401f1085af8d9c12a8bda27375d1f7e Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.568586 4715 generic.go:334] "Generic (PLEG): container finished" podID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerID="7fee4b5921b74233191ea3b7b9453b2f5a0b56d0c5ebd5e8ecc8efa67620a99a" exitCode=0 Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.568774 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" event={"ID":"8ff1994e-54dc-4697-9b2f-402bfb693383","Type":"ContainerDied","Data":"7fee4b5921b74233191ea3b7b9453b2f5a0b56d0c5ebd5e8ecc8efa67620a99a"} Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.568936 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" event={"ID":"8ff1994e-54dc-4697-9b2f-402bfb693383","Type":"ContainerStarted","Data":"cc1f3dce916099d4ba5c7c297c3df73c3f0c3ac63010bc80ba2e02ea5ae4dc5e"} Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.571989 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"a8ac7da1189acfcfc330e1c55794012f8d21cfae15b86b4f788d746bf60937a7"} Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.572030 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"9c27fa98c585c18df0be076f09b22f174292b2bd2e4a179f8cd7ea3d64f1f74a"} Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.576981 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" event={"ID":"552e79b9-1232-48cb-8f68-02488b8d34e7","Type":"ContainerStarted","Data":"96e846f22368bd47c5265100c18d43d7b9cc4b3a70e58a090b066b7dbbc2461f"} Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.577026 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" event={"ID":"552e79b9-1232-48cb-8f68-02488b8d34e7","Type":"ContainerStarted","Data":"e89bb8543537f623d6ad57f65ee54b6610873c4b43d4913a4e0bb5b2df0fd9c2"} Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.577039 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" event={"ID":"552e79b9-1232-48cb-8f68-02488b8d34e7","Type":"ContainerStarted","Data":"9903ea44e7f5a645d518290efdc04e4a2401f1085af8d9c12a8bda27375d1f7e"} Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.577290 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.577315 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:17 crc kubenswrapper[4715]: I1125 12:25:17.609169 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" podStartSLOduration=17.609152392 podStartE2EDuration="17.609152392s" podCreationTimestamp="2025-11-25 12:25:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:25:17.606273847 +0000 UTC m=+968.113776888" watchObservedRunningTime="2025-11-25 12:25:17.609152392 +0000 UTC m=+968.116655413" Nov 25 12:25:18 crc kubenswrapper[4715]: I1125 12:25:18.591491 4715 generic.go:334] "Generic (PLEG): container finished" podID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerID="e16c22542adee6aec27c8db6ca635c60aeadb43220def02a151ac8b2b120247c" exitCode=0 Nov 25 12:25:18 crc kubenswrapper[4715]: I1125 12:25:18.591564 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" event={"ID":"8ff1994e-54dc-4697-9b2f-402bfb693383","Type":"ContainerDied","Data":"e16c22542adee6aec27c8db6ca635c60aeadb43220def02a151ac8b2b120247c"} Nov 25 12:25:18 crc kubenswrapper[4715]: I1125 12:25:18.606559 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"304a0fcab0a47da851c7c328d0ef39c104f864c5cab17780be350bce54dc904a"} Nov 25 12:25:18 crc kubenswrapper[4715]: I1125 12:25:18.606602 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"4ea1c3547f56cb0a07e978953dc99a6a7a7fd92bbd33ed0107c00681c3a4ee79"} Nov 25 12:25:19 crc kubenswrapper[4715]: I1125 12:25:19.621373 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"3d789e85d7e057094ad196b119b5e2b818fffcb9145c05be668ec431091cabd1"} Nov 25 12:25:19 crc kubenswrapper[4715]: I1125 12:25:19.622264 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"3c918e840484c55e79f35c95761af33cdba922179d798ad516e327bc4774bfe8"} Nov 25 12:25:19 crc kubenswrapper[4715]: I1125 12:25:19.622446 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"442a4857911ca0f5eff01ce1fe2aa5265a0c31d36dbc1499b4bb84ddf50fcfc4"} Nov 25 12:25:19 crc kubenswrapper[4715]: I1125 12:25:19.624245 4715 generic.go:334] "Generic (PLEG): container finished" podID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerID="576d0172560c203849dd4792ba63de3c8337dadafd25a2bae7bc9174ba8bf8a6" exitCode=0 Nov 25 12:25:19 crc kubenswrapper[4715]: I1125 12:25:19.624303 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" event={"ID":"8ff1994e-54dc-4697-9b2f-402bfb693383","Type":"ContainerDied","Data":"576d0172560c203849dd4792ba63de3c8337dadafd25a2bae7bc9174ba8bf8a6"} Nov 25 12:25:20 crc kubenswrapper[4715]: I1125 12:25:20.634556 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"e8c20d57e408c91a77326a2a8675710d63373c6df1546a131e67be859e938477"} Nov 25 12:25:20 crc kubenswrapper[4715]: I1125 12:25:20.958597 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.091166 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-util\") pod \"8ff1994e-54dc-4697-9b2f-402bfb693383\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.091268 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-bundle\") pod \"8ff1994e-54dc-4697-9b2f-402bfb693383\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.091406 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbcpr\" (UniqueName: \"kubernetes.io/projected/8ff1994e-54dc-4697-9b2f-402bfb693383-kube-api-access-xbcpr\") pod \"8ff1994e-54dc-4697-9b2f-402bfb693383\" (UID: \"8ff1994e-54dc-4697-9b2f-402bfb693383\") " Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.092168 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-bundle" (OuterVolumeSpecName: "bundle") pod "8ff1994e-54dc-4697-9b2f-402bfb693383" (UID: "8ff1994e-54dc-4697-9b2f-402bfb693383"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.096992 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ff1994e-54dc-4697-9b2f-402bfb693383-kube-api-access-xbcpr" (OuterVolumeSpecName: "kube-api-access-xbcpr") pod "8ff1994e-54dc-4697-9b2f-402bfb693383" (UID: "8ff1994e-54dc-4697-9b2f-402bfb693383"). InnerVolumeSpecName "kube-api-access-xbcpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.106963 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-util" (OuterVolumeSpecName: "util") pod "8ff1994e-54dc-4697-9b2f-402bfb693383" (UID: "8ff1994e-54dc-4697-9b2f-402bfb693383"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.192546 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbcpr\" (UniqueName: \"kubernetes.io/projected/8ff1994e-54dc-4697-9b2f-402bfb693383-kube-api-access-xbcpr\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.192584 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-util\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.192597 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8ff1994e-54dc-4697-9b2f-402bfb693383-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.658110 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"02fb70738920f4cd0ce6d9c1399b1cb7a3eba9d296ff8ca922c42a9d76dfe36b"} Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.666531 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" event={"ID":"8ff1994e-54dc-4697-9b2f-402bfb693383","Type":"ContainerDied","Data":"cc1f3dce916099d4ba5c7c297c3df73c3f0c3ac63010bc80ba2e02ea5ae4dc5e"} Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.666572 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc1f3dce916099d4ba5c7c297c3df73c3f0c3ac63010bc80ba2e02ea5ae4dc5e" Nov 25 12:25:21 crc kubenswrapper[4715]: I1125 12:25:21.666639 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs" Nov 25 12:25:22 crc kubenswrapper[4715]: I1125 12:25:22.680151 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"932022ea4363915c8a857d17744c296deace9e96d23c67a2f0d72d2e2b4593a6"} Nov 25 12:25:22 crc kubenswrapper[4715]: I1125 12:25:22.680470 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"921bfa24e50b593fa54091c88ec254a64337ae8282ac7025335faac84447e848"} Nov 25 12:25:22 crc kubenswrapper[4715]: I1125 12:25:22.680483 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"d41d3fe87731487092ab56c490d7ad02d8567aa3813db343014ee1f1a1025b61"} Nov 25 12:25:22 crc kubenswrapper[4715]: I1125 12:25:22.680492 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"8851042cbc389f207ad0a368c86becad750da7cca9dd7c21539abffa78fe9c4d"} Nov 25 12:25:22 crc kubenswrapper[4715]: I1125 12:25:22.680501 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"623eec66831ff09444cd12ab3cd6be1c114896563eef4d664eb5bcd625c0a8f5"} Nov 25 12:25:22 crc kubenswrapper[4715]: I1125 12:25:22.680511 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"edd3d91a-8a99-4a84-880f-75c3968b264e","Type":"ContainerStarted","Data":"59440ad9903e9d8981898496594777597ebdbf6f7169b645acf5626c6da22930"} Nov 25 12:25:22 crc kubenswrapper[4715]: I1125 12:25:22.724667 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=19.179637136 podStartE2EDuration="24.724635041s" podCreationTimestamp="2025-11-25 12:24:58 +0000 UTC" firstStartedPulling="2025-11-25 12:25:15.921831651 +0000 UTC m=+966.429334672" lastFinishedPulling="2025-11-25 12:25:21.466829556 +0000 UTC m=+971.974332577" observedRunningTime="2025-11-25 12:25:22.72304252 +0000 UTC m=+973.230545541" watchObservedRunningTime="2025-11-25 12:25:22.724635041 +0000 UTC m=+973.232138062" Nov 25 12:25:26 crc kubenswrapper[4715]: I1125 12:25:26.482421 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:26 crc kubenswrapper[4715]: I1125 12:25:26.482768 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-jlm2n" Nov 25 12:25:26 crc kubenswrapper[4715]: I1125 12:25:26.944368 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-858pn"] Nov 25 12:25:26 crc kubenswrapper[4715]: E1125 12:25:26.944698 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerName="util" Nov 25 12:25:26 crc kubenswrapper[4715]: I1125 12:25:26.944715 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerName="util" Nov 25 12:25:26 crc kubenswrapper[4715]: E1125 12:25:26.944726 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerName="extract" Nov 25 12:25:26 crc kubenswrapper[4715]: I1125 12:25:26.944748 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerName="extract" Nov 25 12:25:26 crc kubenswrapper[4715]: E1125 12:25:26.944759 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerName="pull" Nov 25 12:25:26 crc kubenswrapper[4715]: I1125 12:25:26.944765 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerName="pull" Nov 25 12:25:26 crc kubenswrapper[4715]: I1125 12:25:26.944946 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ff1994e-54dc-4697-9b2f-402bfb693383" containerName="extract" Nov 25 12:25:26 crc kubenswrapper[4715]: I1125 12:25:26.945956 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:26 crc kubenswrapper[4715]: I1125 12:25:26.970949 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-858pn"] Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.079520 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-catalog-content\") pod \"certified-operators-858pn\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.079577 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-utilities\") pod \"certified-operators-858pn\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.079649 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g44b5\" (UniqueName: \"kubernetes.io/projected/da557569-d135-4e0d-aa7a-04949be94d87-kube-api-access-g44b5\") pod \"certified-operators-858pn\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.180832 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-catalog-content\") pod \"certified-operators-858pn\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.180905 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-utilities\") pod \"certified-operators-858pn\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.181552 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g44b5\" (UniqueName: \"kubernetes.io/projected/da557569-d135-4e0d-aa7a-04949be94d87-kube-api-access-g44b5\") pod \"certified-operators-858pn\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.181691 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-catalog-content\") pod \"certified-operators-858pn\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.181742 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-utilities\") pod \"certified-operators-858pn\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.200597 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g44b5\" (UniqueName: \"kubernetes.io/projected/da557569-d135-4e0d-aa7a-04949be94d87-kube-api-access-g44b5\") pod \"certified-operators-858pn\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.262599 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:27 crc kubenswrapper[4715]: I1125 12:25:27.705128 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-858pn"] Nov 25 12:25:28 crc kubenswrapper[4715]: I1125 12:25:28.726970 4715 generic.go:334] "Generic (PLEG): container finished" podID="da557569-d135-4e0d-aa7a-04949be94d87" containerID="9610e3394576c745a22029e06c4764ee8605bae3ccd7999d46d7080582892826" exitCode=0 Nov 25 12:25:28 crc kubenswrapper[4715]: I1125 12:25:28.727052 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-858pn" event={"ID":"da557569-d135-4e0d-aa7a-04949be94d87","Type":"ContainerDied","Data":"9610e3394576c745a22029e06c4764ee8605bae3ccd7999d46d7080582892826"} Nov 25 12:25:28 crc kubenswrapper[4715]: I1125 12:25:28.727270 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-858pn" event={"ID":"da557569-d135-4e0d-aa7a-04949be94d87","Type":"ContainerStarted","Data":"d2960b0c357d565428e03bdd6590e4bb65a2cb0b24f749e548ab34727f384006"} Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.212555 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj"] Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.214221 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.218080 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.218480 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-z2dbf" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.233644 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj"] Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.253283 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5f528c96-c714-485d-8486-0b5f6efa0973-apiservice-cert\") pod \"glance-operator-controller-manager-69cb78548c-2nfnj\" (UID: \"5f528c96-c714-485d-8486-0b5f6efa0973\") " pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.253554 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5f528c96-c714-485d-8486-0b5f6efa0973-webhook-cert\") pod \"glance-operator-controller-manager-69cb78548c-2nfnj\" (UID: \"5f528c96-c714-485d-8486-0b5f6efa0973\") " pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.253707 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdtwg\" (UniqueName: \"kubernetes.io/projected/5f528c96-c714-485d-8486-0b5f6efa0973-kube-api-access-jdtwg\") pod \"glance-operator-controller-manager-69cb78548c-2nfnj\" (UID: \"5f528c96-c714-485d-8486-0b5f6efa0973\") " pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.355012 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5f528c96-c714-485d-8486-0b5f6efa0973-apiservice-cert\") pod \"glance-operator-controller-manager-69cb78548c-2nfnj\" (UID: \"5f528c96-c714-485d-8486-0b5f6efa0973\") " pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.355290 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5f528c96-c714-485d-8486-0b5f6efa0973-webhook-cert\") pod \"glance-operator-controller-manager-69cb78548c-2nfnj\" (UID: \"5f528c96-c714-485d-8486-0b5f6efa0973\") " pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.355465 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdtwg\" (UniqueName: \"kubernetes.io/projected/5f528c96-c714-485d-8486-0b5f6efa0973-kube-api-access-jdtwg\") pod \"glance-operator-controller-manager-69cb78548c-2nfnj\" (UID: \"5f528c96-c714-485d-8486-0b5f6efa0973\") " pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.360817 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5f528c96-c714-485d-8486-0b5f6efa0973-webhook-cert\") pod \"glance-operator-controller-manager-69cb78548c-2nfnj\" (UID: \"5f528c96-c714-485d-8486-0b5f6efa0973\") " pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.360833 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5f528c96-c714-485d-8486-0b5f6efa0973-apiservice-cert\") pod \"glance-operator-controller-manager-69cb78548c-2nfnj\" (UID: \"5f528c96-c714-485d-8486-0b5f6efa0973\") " pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.375402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdtwg\" (UniqueName: \"kubernetes.io/projected/5f528c96-c714-485d-8486-0b5f6efa0973-kube-api-access-jdtwg\") pod \"glance-operator-controller-manager-69cb78548c-2nfnj\" (UID: \"5f528c96-c714-485d-8486-0b5f6efa0973\") " pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.534493 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:32 crc kubenswrapper[4715]: I1125 12:25:32.986506 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj"] Nov 25 12:25:32 crc kubenswrapper[4715]: W1125 12:25:32.991622 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f528c96_c714_485d_8486_0b5f6efa0973.slice/crio-34bc7c902ae518ccc8d83b00f9807fb0adfe5a993ba2dcc8906a0ebae0c30739 WatchSource:0}: Error finding container 34bc7c902ae518ccc8d83b00f9807fb0adfe5a993ba2dcc8906a0ebae0c30739: Status 404 returned error can't find the container with id 34bc7c902ae518ccc8d83b00f9807fb0adfe5a993ba2dcc8906a0ebae0c30739 Nov 25 12:25:33 crc kubenswrapper[4715]: I1125 12:25:33.774398 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" event={"ID":"5f528c96-c714-485d-8486-0b5f6efa0973","Type":"ContainerStarted","Data":"34bc7c902ae518ccc8d83b00f9807fb0adfe5a993ba2dcc8906a0ebae0c30739"} Nov 25 12:25:33 crc kubenswrapper[4715]: I1125 12:25:33.776647 4715 generic.go:334] "Generic (PLEG): container finished" podID="da557569-d135-4e0d-aa7a-04949be94d87" containerID="6bf4d76e9bd613ba1e53149dae4b4891c1e6ca96c5089aa8325e939f56de31c7" exitCode=0 Nov 25 12:25:33 crc kubenswrapper[4715]: I1125 12:25:33.776690 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-858pn" event={"ID":"da557569-d135-4e0d-aa7a-04949be94d87","Type":"ContainerDied","Data":"6bf4d76e9bd613ba1e53149dae4b4891c1e6ca96c5089aa8325e939f56de31c7"} Nov 25 12:25:35 crc kubenswrapper[4715]: I1125 12:25:35.790515 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" event={"ID":"5f528c96-c714-485d-8486-0b5f6efa0973","Type":"ContainerStarted","Data":"4d223a085aafe97f2ad60f48715e8c6444b89345c70f916b7648ef43a71f848c"} Nov 25 12:25:35 crc kubenswrapper[4715]: I1125 12:25:35.791068 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:35 crc kubenswrapper[4715]: I1125 12:25:35.794031 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-858pn" event={"ID":"da557569-d135-4e0d-aa7a-04949be94d87","Type":"ContainerStarted","Data":"ebf718328af2eb94ab169278bf4fb5d0c23d5fe77addc896d0c0e7631b9f00b2"} Nov 25 12:25:35 crc kubenswrapper[4715]: I1125 12:25:35.814128 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" podStartSLOduration=2.031875633 podStartE2EDuration="3.814106134s" podCreationTimestamp="2025-11-25 12:25:32 +0000 UTC" firstStartedPulling="2025-11-25 12:25:32.994287875 +0000 UTC m=+983.501790896" lastFinishedPulling="2025-11-25 12:25:34.776518376 +0000 UTC m=+985.284021397" observedRunningTime="2025-11-25 12:25:35.811887056 +0000 UTC m=+986.319390087" watchObservedRunningTime="2025-11-25 12:25:35.814106134 +0000 UTC m=+986.321609155" Nov 25 12:25:35 crc kubenswrapper[4715]: I1125 12:25:35.831224 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-858pn" podStartSLOduration=3.441103968 podStartE2EDuration="9.831206883s" podCreationTimestamp="2025-11-25 12:25:26 +0000 UTC" firstStartedPulling="2025-11-25 12:25:28.730887317 +0000 UTC m=+979.238390338" lastFinishedPulling="2025-11-25 12:25:35.120990232 +0000 UTC m=+985.628493253" observedRunningTime="2025-11-25 12:25:35.830068913 +0000 UTC m=+986.337571934" watchObservedRunningTime="2025-11-25 12:25:35.831206883 +0000 UTC m=+986.338709904" Nov 25 12:25:37 crc kubenswrapper[4715]: I1125 12:25:37.263044 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:37 crc kubenswrapper[4715]: I1125 12:25:37.263089 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:37 crc kubenswrapper[4715]: I1125 12:25:37.302377 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:42 crc kubenswrapper[4715]: I1125 12:25:42.540075 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-69cb78548c-2nfnj" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.656899 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.658633 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.660489 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-pr9b5" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.661243 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.661524 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.665996 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.675056 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.684274 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-czw8b"] Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.686013 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.692480 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-63a3-account-create-update-lfs2c"] Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.695571 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.698535 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.703508 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-czw8b"] Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.709434 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-63a3-account-create-update-lfs2c"] Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.713919 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-openstack-scripts\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.714018 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/676324c4-624e-44b5-adf0-f22c0e724765-operator-scripts\") pod \"glance-db-create-czw8b\" (UID: \"676324c4-624e-44b5-adf0-f22c0e724765\") " pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.714037 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znzp7\" (UniqueName: \"kubernetes.io/projected/676324c4-624e-44b5-adf0-f22c0e724765-kube-api-access-znzp7\") pod \"glance-db-create-czw8b\" (UID: \"676324c4-624e-44b5-adf0-f22c0e724765\") " pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.714065 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-openstack-config\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.714088 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpc7b\" (UniqueName: \"kubernetes.io/projected/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-kube-api-access-vpc7b\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.714106 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-openstack-config-secret\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.714126 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a82194-e1ef-4c16-ba40-fcbc435a029a-operator-scripts\") pod \"glance-63a3-account-create-update-lfs2c\" (UID: \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\") " pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.714166 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqvrd\" (UniqueName: \"kubernetes.io/projected/b9a82194-e1ef-4c16-ba40-fcbc435a029a-kube-api-access-hqvrd\") pod \"glance-63a3-account-create-update-lfs2c\" (UID: \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\") " pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.815164 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-openstack-config-secret\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.815250 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a82194-e1ef-4c16-ba40-fcbc435a029a-operator-scripts\") pod \"glance-63a3-account-create-update-lfs2c\" (UID: \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\") " pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.815302 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqvrd\" (UniqueName: \"kubernetes.io/projected/b9a82194-e1ef-4c16-ba40-fcbc435a029a-kube-api-access-hqvrd\") pod \"glance-63a3-account-create-update-lfs2c\" (UID: \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\") " pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.815335 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-openstack-scripts\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.815382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/676324c4-624e-44b5-adf0-f22c0e724765-operator-scripts\") pod \"glance-db-create-czw8b\" (UID: \"676324c4-624e-44b5-adf0-f22c0e724765\") " pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.815396 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znzp7\" (UniqueName: \"kubernetes.io/projected/676324c4-624e-44b5-adf0-f22c0e724765-kube-api-access-znzp7\") pod \"glance-db-create-czw8b\" (UID: \"676324c4-624e-44b5-adf0-f22c0e724765\") " pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.815424 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-openstack-config\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.815449 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpc7b\" (UniqueName: \"kubernetes.io/projected/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-kube-api-access-vpc7b\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.816387 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/676324c4-624e-44b5-adf0-f22c0e724765-operator-scripts\") pod \"glance-db-create-czw8b\" (UID: \"676324c4-624e-44b5-adf0-f22c0e724765\") " pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.816838 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a82194-e1ef-4c16-ba40-fcbc435a029a-operator-scripts\") pod \"glance-63a3-account-create-update-lfs2c\" (UID: \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\") " pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.817092 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-openstack-config\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.817476 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-openstack-scripts\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.821597 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-openstack-config-secret\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.832384 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znzp7\" (UniqueName: \"kubernetes.io/projected/676324c4-624e-44b5-adf0-f22c0e724765-kube-api-access-znzp7\") pod \"glance-db-create-czw8b\" (UID: \"676324c4-624e-44b5-adf0-f22c0e724765\") " pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.833066 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqvrd\" (UniqueName: \"kubernetes.io/projected/b9a82194-e1ef-4c16-ba40-fcbc435a029a-kube-api-access-hqvrd\") pod \"glance-63a3-account-create-update-lfs2c\" (UID: \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\") " pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.833166 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpc7b\" (UniqueName: \"kubernetes.io/projected/30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c-kube-api-access-vpc7b\") pod \"openstackclient\" (UID: \"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c\") " pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:43 crc kubenswrapper[4715]: I1125 12:25:43.981534 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 25 12:25:44 crc kubenswrapper[4715]: I1125 12:25:44.007234 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:44 crc kubenswrapper[4715]: I1125 12:25:44.026461 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:44 crc kubenswrapper[4715]: I1125 12:25:44.433219 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 25 12:25:44 crc kubenswrapper[4715]: W1125 12:25:44.440434 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30f5ad0e_dcd3_49b0_a01f_9d26f7b8047c.slice/crio-921c05ecccc58367376d82952bcb2e9a53b7a50858008c2d64a06bd7b0910819 WatchSource:0}: Error finding container 921c05ecccc58367376d82952bcb2e9a53b7a50858008c2d64a06bd7b0910819: Status 404 returned error can't find the container with id 921c05ecccc58367376d82952bcb2e9a53b7a50858008c2d64a06bd7b0910819 Nov 25 12:25:44 crc kubenswrapper[4715]: I1125 12:25:44.521492 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-czw8b"] Nov 25 12:25:44 crc kubenswrapper[4715]: W1125 12:25:44.522479 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod676324c4_624e_44b5_adf0_f22c0e724765.slice/crio-68ec6519c1aa3a07e2e80b96093f6d169b5764fe6ee87d2523f07cfdfe72b94c WatchSource:0}: Error finding container 68ec6519c1aa3a07e2e80b96093f6d169b5764fe6ee87d2523f07cfdfe72b94c: Status 404 returned error can't find the container with id 68ec6519c1aa3a07e2e80b96093f6d169b5764fe6ee87d2523f07cfdfe72b94c Nov 25 12:25:44 crc kubenswrapper[4715]: I1125 12:25:44.527641 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-63a3-account-create-update-lfs2c"] Nov 25 12:25:44 crc kubenswrapper[4715]: W1125 12:25:44.538011 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9a82194_e1ef_4c16_ba40_fcbc435a029a.slice/crio-7c40d56ae470cce1c5ac13dce1c6aaa24ee79e4368c0d4dec448d835ef0ddbfd WatchSource:0}: Error finding container 7c40d56ae470cce1c5ac13dce1c6aaa24ee79e4368c0d4dec448d835ef0ddbfd: Status 404 returned error can't find the container with id 7c40d56ae470cce1c5ac13dce1c6aaa24ee79e4368c0d4dec448d835ef0ddbfd Nov 25 12:25:44 crc kubenswrapper[4715]: I1125 12:25:44.859145 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c","Type":"ContainerStarted","Data":"921c05ecccc58367376d82952bcb2e9a53b7a50858008c2d64a06bd7b0910819"} Nov 25 12:25:44 crc kubenswrapper[4715]: I1125 12:25:44.860494 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" event={"ID":"b9a82194-e1ef-4c16-ba40-fcbc435a029a","Type":"ContainerStarted","Data":"7c40d56ae470cce1c5ac13dce1c6aaa24ee79e4368c0d4dec448d835ef0ddbfd"} Nov 25 12:25:44 crc kubenswrapper[4715]: I1125 12:25:44.861669 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-czw8b" event={"ID":"676324c4-624e-44b5-adf0-f22c0e724765","Type":"ContainerStarted","Data":"68ec6519c1aa3a07e2e80b96093f6d169b5764fe6ee87d2523f07cfdfe72b94c"} Nov 25 12:25:45 crc kubenswrapper[4715]: I1125 12:25:45.872308 4715 generic.go:334] "Generic (PLEG): container finished" podID="b9a82194-e1ef-4c16-ba40-fcbc435a029a" containerID="f0b0ce2cf823279bb05351f89239a50ba7b88fb0afaf058b32afdb9472a841d7" exitCode=0 Nov 25 12:25:45 crc kubenswrapper[4715]: I1125 12:25:45.872765 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" event={"ID":"b9a82194-e1ef-4c16-ba40-fcbc435a029a","Type":"ContainerDied","Data":"f0b0ce2cf823279bb05351f89239a50ba7b88fb0afaf058b32afdb9472a841d7"} Nov 25 12:25:45 crc kubenswrapper[4715]: I1125 12:25:45.875957 4715 generic.go:334] "Generic (PLEG): container finished" podID="676324c4-624e-44b5-adf0-f22c0e724765" containerID="68706bddd9c2ca96efb4927dd5e3280d7ea7228514e773f587a18590558e1974" exitCode=0 Nov 25 12:25:45 crc kubenswrapper[4715]: I1125 12:25:45.876024 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-czw8b" event={"ID":"676324c4-624e-44b5-adf0-f22c0e724765","Type":"ContainerDied","Data":"68706bddd9c2ca96efb4927dd5e3280d7ea7228514e773f587a18590558e1974"} Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.321428 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.325516 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.411854 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.470736 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqvrd\" (UniqueName: \"kubernetes.io/projected/b9a82194-e1ef-4c16-ba40-fcbc435a029a-kube-api-access-hqvrd\") pod \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\" (UID: \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\") " Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.470846 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a82194-e1ef-4c16-ba40-fcbc435a029a-operator-scripts\") pod \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\" (UID: \"b9a82194-e1ef-4c16-ba40-fcbc435a029a\") " Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.472008 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a82194-e1ef-4c16-ba40-fcbc435a029a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b9a82194-e1ef-4c16-ba40-fcbc435a029a" (UID: "b9a82194-e1ef-4c16-ba40-fcbc435a029a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.477054 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a82194-e1ef-4c16-ba40-fcbc435a029a-kube-api-access-hqvrd" (OuterVolumeSpecName: "kube-api-access-hqvrd") pod "b9a82194-e1ef-4c16-ba40-fcbc435a029a" (UID: "b9a82194-e1ef-4c16-ba40-fcbc435a029a"). InnerVolumeSpecName "kube-api-access-hqvrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.572733 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/676324c4-624e-44b5-adf0-f22c0e724765-operator-scripts\") pod \"676324c4-624e-44b5-adf0-f22c0e724765\" (UID: \"676324c4-624e-44b5-adf0-f22c0e724765\") " Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.573113 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znzp7\" (UniqueName: \"kubernetes.io/projected/676324c4-624e-44b5-adf0-f22c0e724765-kube-api-access-znzp7\") pod \"676324c4-624e-44b5-adf0-f22c0e724765\" (UID: \"676324c4-624e-44b5-adf0-f22c0e724765\") " Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.573284 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/676324c4-624e-44b5-adf0-f22c0e724765-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "676324c4-624e-44b5-adf0-f22c0e724765" (UID: "676324c4-624e-44b5-adf0-f22c0e724765"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.573571 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqvrd\" (UniqueName: \"kubernetes.io/projected/b9a82194-e1ef-4c16-ba40-fcbc435a029a-kube-api-access-hqvrd\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.573596 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a82194-e1ef-4c16-ba40-fcbc435a029a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.573610 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/676324c4-624e-44b5-adf0-f22c0e724765-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.576391 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/676324c4-624e-44b5-adf0-f22c0e724765-kube-api-access-znzp7" (OuterVolumeSpecName: "kube-api-access-znzp7") pod "676324c4-624e-44b5-adf0-f22c0e724765" (UID: "676324c4-624e-44b5-adf0-f22c0e724765"). InnerVolumeSpecName "kube-api-access-znzp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.674696 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znzp7\" (UniqueName: \"kubernetes.io/projected/676324c4-624e-44b5-adf0-f22c0e724765-kube-api-access-znzp7\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.894680 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-czw8b" event={"ID":"676324c4-624e-44b5-adf0-f22c0e724765","Type":"ContainerDied","Data":"68ec6519c1aa3a07e2e80b96093f6d169b5764fe6ee87d2523f07cfdfe72b94c"} Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.894723 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68ec6519c1aa3a07e2e80b96093f6d169b5764fe6ee87d2523f07cfdfe72b94c" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.894735 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-czw8b" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.898344 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" event={"ID":"b9a82194-e1ef-4c16-ba40-fcbc435a029a","Type":"ContainerDied","Data":"7c40d56ae470cce1c5ac13dce1c6aaa24ee79e4368c0d4dec448d835ef0ddbfd"} Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.898379 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c40d56ae470cce1c5ac13dce1c6aaa24ee79e4368c0d4dec448d835ef0ddbfd" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.898392 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-63a3-account-create-update-lfs2c" Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.943557 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:25:47 crc kubenswrapper[4715]: I1125 12:25:47.943655 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.786665 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-7xvp9"] Nov 25 12:25:48 crc kubenswrapper[4715]: E1125 12:25:48.786976 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="676324c4-624e-44b5-adf0-f22c0e724765" containerName="mariadb-database-create" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.786990 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="676324c4-624e-44b5-adf0-f22c0e724765" containerName="mariadb-database-create" Nov 25 12:25:48 crc kubenswrapper[4715]: E1125 12:25:48.787009 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a82194-e1ef-4c16-ba40-fcbc435a029a" containerName="mariadb-account-create-update" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.787015 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a82194-e1ef-4c16-ba40-fcbc435a029a" containerName="mariadb-account-create-update" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.787134 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a82194-e1ef-4c16-ba40-fcbc435a029a" containerName="mariadb-account-create-update" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.787148 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="676324c4-624e-44b5-adf0-f22c0e724765" containerName="mariadb-database-create" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.787602 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.791178 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.791484 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-7gbkp" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.799035 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-7xvp9"] Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.992414 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5snz\" (UniqueName: \"kubernetes.io/projected/31de582a-c3fe-412a-a6a6-fde502f793b1-kube-api-access-c5snz\") pod \"glance-db-sync-7xvp9\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.992484 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-db-sync-config-data\") pod \"glance-db-sync-7xvp9\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:48 crc kubenswrapper[4715]: I1125 12:25:48.992516 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-config-data\") pod \"glance-db-sync-7xvp9\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:49 crc kubenswrapper[4715]: I1125 12:25:49.093511 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5snz\" (UniqueName: \"kubernetes.io/projected/31de582a-c3fe-412a-a6a6-fde502f793b1-kube-api-access-c5snz\") pod \"glance-db-sync-7xvp9\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:49 crc kubenswrapper[4715]: I1125 12:25:49.094786 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-db-sync-config-data\") pod \"glance-db-sync-7xvp9\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:49 crc kubenswrapper[4715]: I1125 12:25:49.094812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-config-data\") pod \"glance-db-sync-7xvp9\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:49 crc kubenswrapper[4715]: I1125 12:25:49.100117 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-config-data\") pod \"glance-db-sync-7xvp9\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:49 crc kubenswrapper[4715]: I1125 12:25:49.100158 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-db-sync-config-data\") pod \"glance-db-sync-7xvp9\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:49 crc kubenswrapper[4715]: I1125 12:25:49.109975 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5snz\" (UniqueName: \"kubernetes.io/projected/31de582a-c3fe-412a-a6a6-fde502f793b1-kube-api-access-c5snz\") pod \"glance-db-sync-7xvp9\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:49 crc kubenswrapper[4715]: I1125 12:25:49.404118 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:25:50 crc kubenswrapper[4715]: I1125 12:25:50.909880 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-858pn"] Nov 25 12:25:50 crc kubenswrapper[4715]: I1125 12:25:50.911085 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-858pn" podUID="da557569-d135-4e0d-aa7a-04949be94d87" containerName="registry-server" containerID="cri-o://ebf718328af2eb94ab169278bf4fb5d0c23d5fe77addc896d0c0e7631b9f00b2" gracePeriod=2 Nov 25 12:25:51 crc kubenswrapper[4715]: I1125 12:25:51.946296 4715 generic.go:334] "Generic (PLEG): container finished" podID="da557569-d135-4e0d-aa7a-04949be94d87" containerID="ebf718328af2eb94ab169278bf4fb5d0c23d5fe77addc896d0c0e7631b9f00b2" exitCode=0 Nov 25 12:25:51 crc kubenswrapper[4715]: I1125 12:25:51.946341 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-858pn" event={"ID":"da557569-d135-4e0d-aa7a-04949be94d87","Type":"ContainerDied","Data":"ebf718328af2eb94ab169278bf4fb5d0c23d5fe77addc896d0c0e7631b9f00b2"} Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.771727 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.945072 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-7xvp9"] Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.948980 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g44b5\" (UniqueName: \"kubernetes.io/projected/da557569-d135-4e0d-aa7a-04949be94d87-kube-api-access-g44b5\") pod \"da557569-d135-4e0d-aa7a-04949be94d87\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.949039 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-catalog-content\") pod \"da557569-d135-4e0d-aa7a-04949be94d87\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.949194 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-utilities\") pod \"da557569-d135-4e0d-aa7a-04949be94d87\" (UID: \"da557569-d135-4e0d-aa7a-04949be94d87\") " Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.949978 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-utilities" (OuterVolumeSpecName: "utilities") pod "da557569-d135-4e0d-aa7a-04949be94d87" (UID: "da557569-d135-4e0d-aa7a-04949be94d87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.953630 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da557569-d135-4e0d-aa7a-04949be94d87-kube-api-access-g44b5" (OuterVolumeSpecName: "kube-api-access-g44b5") pod "da557569-d135-4e0d-aa7a-04949be94d87" (UID: "da557569-d135-4e0d-aa7a-04949be94d87"). InnerVolumeSpecName "kube-api-access-g44b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.956204 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-858pn" event={"ID":"da557569-d135-4e0d-aa7a-04949be94d87","Type":"ContainerDied","Data":"d2960b0c357d565428e03bdd6590e4bb65a2cb0b24f749e548ab34727f384006"} Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.956286 4715 scope.go:117] "RemoveContainer" containerID="ebf718328af2eb94ab169278bf4fb5d0c23d5fe77addc896d0c0e7631b9f00b2" Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.956526 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-858pn" Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.959591 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c","Type":"ContainerStarted","Data":"b1544947c33394c0ac871f1d89bf4f71b4a66c77bd8c7ae714f3d0e35bf0ccb7"} Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.972677 4715 scope.go:117] "RemoveContainer" containerID="6bf4d76e9bd613ba1e53149dae4b4891c1e6ca96c5089aa8325e939f56de31c7" Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.978987 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.80186858 podStartE2EDuration="9.978968452s" podCreationTimestamp="2025-11-25 12:25:43 +0000 UTC" firstStartedPulling="2025-11-25 12:25:44.443813148 +0000 UTC m=+994.951316169" lastFinishedPulling="2025-11-25 12:25:52.62091302 +0000 UTC m=+1003.128416041" observedRunningTime="2025-11-25 12:25:52.973040997 +0000 UTC m=+1003.480544018" watchObservedRunningTime="2025-11-25 12:25:52.978968452 +0000 UTC m=+1003.486471473" Nov 25 12:25:52 crc kubenswrapper[4715]: I1125 12:25:52.988820 4715 scope.go:117] "RemoveContainer" containerID="9610e3394576c745a22029e06c4764ee8605bae3ccd7999d46d7080582892826" Nov 25 12:25:53 crc kubenswrapper[4715]: I1125 12:25:53.002050 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da557569-d135-4e0d-aa7a-04949be94d87" (UID: "da557569-d135-4e0d-aa7a-04949be94d87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:25:53 crc kubenswrapper[4715]: I1125 12:25:53.051274 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:53 crc kubenswrapper[4715]: I1125 12:25:53.051337 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g44b5\" (UniqueName: \"kubernetes.io/projected/da557569-d135-4e0d-aa7a-04949be94d87-kube-api-access-g44b5\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:53 crc kubenswrapper[4715]: I1125 12:25:53.051354 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da557569-d135-4e0d-aa7a-04949be94d87-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:25:53 crc kubenswrapper[4715]: I1125 12:25:53.287397 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-858pn"] Nov 25 12:25:53 crc kubenswrapper[4715]: I1125 12:25:53.298669 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-858pn"] Nov 25 12:25:53 crc kubenswrapper[4715]: I1125 12:25:53.971151 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-7xvp9" event={"ID":"31de582a-c3fe-412a-a6a6-fde502f793b1","Type":"ContainerStarted","Data":"1c55f39990cae3df5a81fcea53465c1268a80e40d725fb30288883c7bbb74bfc"} Nov 25 12:25:54 crc kubenswrapper[4715]: I1125 12:25:54.703559 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da557569-d135-4e0d-aa7a-04949be94d87" path="/var/lib/kubelet/pods/da557569-d135-4e0d-aa7a-04949be94d87/volumes" Nov 25 12:26:06 crc kubenswrapper[4715]: I1125 12:26:06.060811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-7xvp9" event={"ID":"31de582a-c3fe-412a-a6a6-fde502f793b1","Type":"ContainerStarted","Data":"f653bc3cb656cf3aa2e877aa07917f272672cc5b2803983aade68543e112a904"} Nov 25 12:26:06 crc kubenswrapper[4715]: I1125 12:26:06.085941 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-7xvp9" podStartSLOduration=5.981753738 podStartE2EDuration="18.085922815s" podCreationTimestamp="2025-11-25 12:25:48 +0000 UTC" firstStartedPulling="2025-11-25 12:25:52.954060639 +0000 UTC m=+1003.461563660" lastFinishedPulling="2025-11-25 12:26:05.058229716 +0000 UTC m=+1015.565732737" observedRunningTime="2025-11-25 12:26:06.080921533 +0000 UTC m=+1016.588424564" watchObservedRunningTime="2025-11-25 12:26:06.085922815 +0000 UTC m=+1016.593425846" Nov 25 12:26:14 crc kubenswrapper[4715]: I1125 12:26:14.121613 4715 generic.go:334] "Generic (PLEG): container finished" podID="31de582a-c3fe-412a-a6a6-fde502f793b1" containerID="f653bc3cb656cf3aa2e877aa07917f272672cc5b2803983aade68543e112a904" exitCode=0 Nov 25 12:26:14 crc kubenswrapper[4715]: I1125 12:26:14.121664 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-7xvp9" event={"ID":"31de582a-c3fe-412a-a6a6-fde502f793b1","Type":"ContainerDied","Data":"f653bc3cb656cf3aa2e877aa07917f272672cc5b2803983aade68543e112a904"} Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.419795 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.590015 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5snz\" (UniqueName: \"kubernetes.io/projected/31de582a-c3fe-412a-a6a6-fde502f793b1-kube-api-access-c5snz\") pod \"31de582a-c3fe-412a-a6a6-fde502f793b1\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.590131 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-config-data\") pod \"31de582a-c3fe-412a-a6a6-fde502f793b1\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.590221 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-db-sync-config-data\") pod \"31de582a-c3fe-412a-a6a6-fde502f793b1\" (UID: \"31de582a-c3fe-412a-a6a6-fde502f793b1\") " Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.596356 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31de582a-c3fe-412a-a6a6-fde502f793b1-kube-api-access-c5snz" (OuterVolumeSpecName: "kube-api-access-c5snz") pod "31de582a-c3fe-412a-a6a6-fde502f793b1" (UID: "31de582a-c3fe-412a-a6a6-fde502f793b1"). InnerVolumeSpecName "kube-api-access-c5snz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.597242 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "31de582a-c3fe-412a-a6a6-fde502f793b1" (UID: "31de582a-c3fe-412a-a6a6-fde502f793b1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.634621 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-config-data" (OuterVolumeSpecName: "config-data") pod "31de582a-c3fe-412a-a6a6-fde502f793b1" (UID: "31de582a-c3fe-412a-a6a6-fde502f793b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.692120 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5snz\" (UniqueName: \"kubernetes.io/projected/31de582a-c3fe-412a-a6a6-fde502f793b1-kube-api-access-c5snz\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.692163 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:15 crc kubenswrapper[4715]: I1125 12:26:15.692177 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31de582a-c3fe-412a-a6a6-fde502f793b1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:16 crc kubenswrapper[4715]: I1125 12:26:16.136980 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-7xvp9" event={"ID":"31de582a-c3fe-412a-a6a6-fde502f793b1","Type":"ContainerDied","Data":"1c55f39990cae3df5a81fcea53465c1268a80e40d725fb30288883c7bbb74bfc"} Nov 25 12:26:16 crc kubenswrapper[4715]: I1125 12:26:16.137024 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c55f39990cae3df5a81fcea53465c1268a80e40d725fb30288883c7bbb74bfc" Nov 25 12:26:16 crc kubenswrapper[4715]: I1125 12:26:16.137063 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-7xvp9" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.364635 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 12:26:17 crc kubenswrapper[4715]: E1125 12:26:17.365249 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da557569-d135-4e0d-aa7a-04949be94d87" containerName="registry-server" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.365262 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="da557569-d135-4e0d-aa7a-04949be94d87" containerName="registry-server" Nov 25 12:26:17 crc kubenswrapper[4715]: E1125 12:26:17.365275 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da557569-d135-4e0d-aa7a-04949be94d87" containerName="extract-utilities" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.365281 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="da557569-d135-4e0d-aa7a-04949be94d87" containerName="extract-utilities" Nov 25 12:26:17 crc kubenswrapper[4715]: E1125 12:26:17.365293 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31de582a-c3fe-412a-a6a6-fde502f793b1" containerName="glance-db-sync" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.365299 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="31de582a-c3fe-412a-a6a6-fde502f793b1" containerName="glance-db-sync" Nov 25 12:26:17 crc kubenswrapper[4715]: E1125 12:26:17.365314 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da557569-d135-4e0d-aa7a-04949be94d87" containerName="extract-content" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.365320 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="da557569-d135-4e0d-aa7a-04949be94d87" containerName="extract-content" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.365448 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="da557569-d135-4e0d-aa7a-04949be94d87" containerName="registry-server" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.365459 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="31de582a-c3fe-412a-a6a6-fde502f793b1" containerName="glance-db-sync" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.366173 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.367999 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.368412 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-7gbkp" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.368559 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.382664 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.516965 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517034 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-logs\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517066 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517119 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-run\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517163 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-dev\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517213 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-lib-modules\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517239 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-scripts\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517290 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-httpd-run\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517315 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-config-data\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517351 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-nvme\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517385 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvghk\" (UniqueName: \"kubernetes.io/projected/d4b54693-6831-4fb0-b631-7950e9a9234b-kube-api-access-lvghk\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517420 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517455 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-sys\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.517489 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.544286 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.545764 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.556582 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619266 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-httpd-run\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619307 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-config-data\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619336 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-nvme\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619361 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvghk\" (UniqueName: \"kubernetes.io/projected/d4b54693-6831-4fb0-b631-7950e9a9234b-kube-api-access-lvghk\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619384 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619413 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-sys\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619468 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619495 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619532 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-logs\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619556 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619531 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-sys\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619604 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-run\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619576 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-run\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619630 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-dev\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-lib-modules\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619674 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-scripts\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619764 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-lib-modules\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619791 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619804 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619831 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619755 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-dev\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.619943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-httpd-run\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.620021 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-logs\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.620483 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-nvme\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.623987 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-scripts\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.625013 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-config-data\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.638834 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvghk\" (UniqueName: \"kubernetes.io/projected/d4b54693-6831-4fb0-b631-7950e9a9234b-kube-api-access-lvghk\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.647800 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.652392 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.689995 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721340 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-httpd-run\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721411 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-logs\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721460 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721510 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-lib-modules\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721586 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721606 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlzl9\" (UniqueName: \"kubernetes.io/projected/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-kube-api-access-dlzl9\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721630 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-sys\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721758 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-config-data\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721838 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-scripts\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721856 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721885 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-run\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.721963 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-dev\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823468 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-dev\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823719 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-httpd-run\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823754 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-logs\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823792 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823820 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-lib-modules\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823840 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlzl9\" (UniqueName: \"kubernetes.io/projected/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-kube-api-access-dlzl9\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823855 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823875 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-sys\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823913 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.823976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824001 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824051 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-config-data\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824077 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824068 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824093 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-scripts\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824153 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-run\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824279 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-run\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824533 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-dev\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824597 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824623 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-logs\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824011 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-lib-modules\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824793 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824875 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-sys\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.824894 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.825007 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-httpd-run\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.829592 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-scripts\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.832312 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-config-data\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.851985 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlzl9\" (UniqueName: \"kubernetes.io/projected/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-kube-api-access-dlzl9\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.856060 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.866397 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.943841 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:26:17 crc kubenswrapper[4715]: I1125 12:26:17.943903 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:26:18 crc kubenswrapper[4715]: I1125 12:26:18.158840 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 12:26:18 crc kubenswrapper[4715]: I1125 12:26:18.161270 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:18 crc kubenswrapper[4715]: I1125 12:26:18.726543 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:26:18 crc kubenswrapper[4715]: W1125 12:26:18.738110 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7811654a_cd03_48f4_bb9d_1bd6cbf2094c.slice/crio-7277e04b8437da60871a674a9f50f9d50142e84d8922e277d2b8b3c71c6ec113 WatchSource:0}: Error finding container 7277e04b8437da60871a674a9f50f9d50142e84d8922e277d2b8b3c71c6ec113: Status 404 returned error can't find the container with id 7277e04b8437da60871a674a9f50f9d50142e84d8922e277d2b8b3c71c6ec113 Nov 25 12:26:19 crc kubenswrapper[4715]: I1125 12:26:19.161061 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"d4b54693-6831-4fb0-b631-7950e9a9234b","Type":"ContainerStarted","Data":"a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd"} Nov 25 12:26:19 crc kubenswrapper[4715]: I1125 12:26:19.161605 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"d4b54693-6831-4fb0-b631-7950e9a9234b","Type":"ContainerStarted","Data":"8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0"} Nov 25 12:26:19 crc kubenswrapper[4715]: I1125 12:26:19.161617 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"d4b54693-6831-4fb0-b631-7950e9a9234b","Type":"ContainerStarted","Data":"d43d83f1f58e59461b4fa33ac1ce2a190f98bb891b109ce967d0b231a050b79c"} Nov 25 12:26:19 crc kubenswrapper[4715]: I1125 12:26:19.163307 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7811654a-cd03-48f4-bb9d-1bd6cbf2094c","Type":"ContainerStarted","Data":"48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca"} Nov 25 12:26:19 crc kubenswrapper[4715]: I1125 12:26:19.163337 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7811654a-cd03-48f4-bb9d-1bd6cbf2094c","Type":"ContainerStarted","Data":"87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1"} Nov 25 12:26:19 crc kubenswrapper[4715]: I1125 12:26:19.163349 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7811654a-cd03-48f4-bb9d-1bd6cbf2094c","Type":"ContainerStarted","Data":"7277e04b8437da60871a674a9f50f9d50142e84d8922e277d2b8b3c71c6ec113"} Nov 25 12:26:19 crc kubenswrapper[4715]: I1125 12:26:19.188240 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.188221644 podStartE2EDuration="2.188221644s" podCreationTimestamp="2025-11-25 12:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:26:19.181741684 +0000 UTC m=+1029.689244725" watchObservedRunningTime="2025-11-25 12:26:19.188221644 +0000 UTC m=+1029.695724665" Nov 25 12:26:19 crc kubenswrapper[4715]: I1125 12:26:19.202632 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.202616211 podStartE2EDuration="3.202616211s" podCreationTimestamp="2025-11-25 12:26:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:26:19.200080644 +0000 UTC m=+1029.707583665" watchObservedRunningTime="2025-11-25 12:26:19.202616211 +0000 UTC m=+1029.710119232" Nov 25 12:26:27 crc kubenswrapper[4715]: I1125 12:26:27.691121 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:27 crc kubenswrapper[4715]: I1125 12:26:27.693211 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:27 crc kubenswrapper[4715]: I1125 12:26:27.733700 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:27 crc kubenswrapper[4715]: I1125 12:26:27.774697 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:28 crc kubenswrapper[4715]: I1125 12:26:28.161940 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:28 crc kubenswrapper[4715]: I1125 12:26:28.162019 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:28 crc kubenswrapper[4715]: I1125 12:26:28.186053 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:28 crc kubenswrapper[4715]: I1125 12:26:28.220348 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:28 crc kubenswrapper[4715]: I1125 12:26:28.235761 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:28 crc kubenswrapper[4715]: I1125 12:26:28.236041 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:28 crc kubenswrapper[4715]: I1125 12:26:28.236152 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:28 crc kubenswrapper[4715]: I1125 12:26:28.236278 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:30 crc kubenswrapper[4715]: I1125 12:26:30.248131 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 12:26:30 crc kubenswrapper[4715]: I1125 12:26:30.248415 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 12:26:30 crc kubenswrapper[4715]: I1125 12:26:30.248130 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 12:26:30 crc kubenswrapper[4715]: I1125 12:26:30.248653 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 12:26:30 crc kubenswrapper[4715]: I1125 12:26:30.439481 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:30 crc kubenswrapper[4715]: I1125 12:26:30.455677 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:30 crc kubenswrapper[4715]: I1125 12:26:30.456259 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:30 crc kubenswrapper[4715]: I1125 12:26:30.535983 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:26:30 crc kubenswrapper[4715]: I1125 12:26:30.607218 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:26:32 crc kubenswrapper[4715]: I1125 12:26:32.262166 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-log" containerID="cri-o://87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1" gracePeriod=30 Nov 25 12:26:32 crc kubenswrapper[4715]: I1125 12:26:32.262286 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-httpd" containerID="cri-o://48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca" gracePeriod=30 Nov 25 12:26:32 crc kubenswrapper[4715]: I1125 12:26:32.269147 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": EOF" Nov 25 12:26:32 crc kubenswrapper[4715]: I1125 12:26:32.270161 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.101:9292/healthcheck\": EOF" Nov 25 12:26:33 crc kubenswrapper[4715]: I1125 12:26:33.270414 4715 generic.go:334] "Generic (PLEG): container finished" podID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerID="87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1" exitCode=143 Nov 25 12:26:33 crc kubenswrapper[4715]: I1125 12:26:33.270456 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7811654a-cd03-48f4-bb9d-1bd6cbf2094c","Type":"ContainerDied","Data":"87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1"} Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.060979 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138072 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-httpd-run\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138140 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-nvme\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138173 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-logs\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138215 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-var-locks-brick\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138245 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-dev\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138281 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138276 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138307 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-config-data\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138324 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-dev" (OuterVolumeSpecName: "dev") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138331 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-sys\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138376 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlzl9\" (UniqueName: \"kubernetes.io/projected/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-kube-api-access-dlzl9\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138402 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-run\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138436 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138463 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-scripts\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138503 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-iscsi\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138529 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-lib-modules\") pod \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\" (UID: \"7811654a-cd03-48f4-bb9d-1bd6cbf2094c\") " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138552 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-sys" (OuterVolumeSpecName: "sys") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138912 4715 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138929 4715 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-dev\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138939 4715 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-sys\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138635 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138975 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138969 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.138995 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.139019 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-run" (OuterVolumeSpecName: "run") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.143460 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-logs" (OuterVolumeSpecName: "logs") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.147292 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.147366 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.149284 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-scripts" (OuterVolumeSpecName: "scripts") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.149402 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-kube-api-access-dlzl9" (OuterVolumeSpecName: "kube-api-access-dlzl9") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "kube-api-access-dlzl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.186157 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-config-data" (OuterVolumeSpecName: "config-data") pod "7811654a-cd03-48f4-bb9d-1bd6cbf2094c" (UID: "7811654a-cd03-48f4-bb9d-1bd6cbf2094c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240570 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlzl9\" (UniqueName: \"kubernetes.io/projected/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-kube-api-access-dlzl9\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240618 4715 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240661 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240676 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240689 4715 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240700 4715 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240711 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240722 4715 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240731 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-logs\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240752 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.240765 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7811654a-cd03-48f4-bb9d-1bd6cbf2094c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.252490 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.257908 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.300131 4715 generic.go:334] "Generic (PLEG): container finished" podID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerID="48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca" exitCode=0 Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.300196 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7811654a-cd03-48f4-bb9d-1bd6cbf2094c","Type":"ContainerDied","Data":"48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca"} Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.300232 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7811654a-cd03-48f4-bb9d-1bd6cbf2094c","Type":"ContainerDied","Data":"7277e04b8437da60871a674a9f50f9d50142e84d8922e277d2b8b3c71c6ec113"} Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.300254 4715 scope.go:117] "RemoveContainer" containerID="48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.300248 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.322559 4715 scope.go:117] "RemoveContainer" containerID="87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.339406 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.341850 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.341894 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.343860 4715 scope.go:117] "RemoveContainer" containerID="48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca" Nov 25 12:26:37 crc kubenswrapper[4715]: E1125 12:26:37.344253 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca\": container with ID starting with 48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca not found: ID does not exist" containerID="48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.344332 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca"} err="failed to get container status \"48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca\": rpc error: code = NotFound desc = could not find container \"48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca\": container with ID starting with 48856deeecdd75ecc56d25a61fb712cc9e0e9ca2379f4c572ebc162c928056ca not found: ID does not exist" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.344376 4715 scope.go:117] "RemoveContainer" containerID="87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1" Nov 25 12:26:37 crc kubenswrapper[4715]: E1125 12:26:37.344917 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1\": container with ID starting with 87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1 not found: ID does not exist" containerID="87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.344939 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1"} err="failed to get container status \"87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1\": rpc error: code = NotFound desc = could not find container \"87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1\": container with ID starting with 87f4e8b8b0915082958fc967cf6afd3bb65fb22e899414807439fd5ea9d893c1 not found: ID does not exist" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.346974 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.366362 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:26:37 crc kubenswrapper[4715]: E1125 12:26:37.366754 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-httpd" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.366784 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-httpd" Nov 25 12:26:37 crc kubenswrapper[4715]: E1125 12:26:37.366831 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-log" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.366839 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-log" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.366989 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-log" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.367019 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" containerName="glance-httpd" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.368154 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.380423 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443116 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-httpd-run\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443219 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjc56\" (UniqueName: \"kubernetes.io/projected/8167972f-8b24-4c5d-ba63-fe81885b0e8c-kube-api-access-zjc56\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443241 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-lib-modules\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-scripts\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443312 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-config-data\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443370 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-logs\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443494 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-sys\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443565 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-nvme\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443604 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443633 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-run\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443682 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-dev\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443716 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443734 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.443845 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546383 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-nvme\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546483 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-run\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546513 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-dev\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546558 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546596 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546623 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-httpd-run\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546676 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjc56\" (UniqueName: \"kubernetes.io/projected/8167972f-8b24-4c5d-ba63-fe81885b0e8c-kube-api-access-zjc56\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546699 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-lib-modules\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546721 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-scripts\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546781 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-config-data\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546821 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-logs\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546847 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-sys\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546932 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-sys\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.546995 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-nvme\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.547097 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.549217 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.549279 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.549278 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.549295 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-run\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.549305 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-lib-modules\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.549315 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-dev\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.549692 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-logs\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.550014 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-httpd-run\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.554411 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-scripts\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.558699 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-config-data\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.571439 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.573055 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjc56\" (UniqueName: \"kubernetes.io/projected/8167972f-8b24-4c5d-ba63-fe81885b0e8c-kube-api-access-zjc56\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.576153 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:37 crc kubenswrapper[4715]: I1125 12:26:37.686268 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:38 crc kubenswrapper[4715]: I1125 12:26:38.127685 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:26:38 crc kubenswrapper[4715]: I1125 12:26:38.310696 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8167972f-8b24-4c5d-ba63-fe81885b0e8c","Type":"ContainerStarted","Data":"82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336"} Nov 25 12:26:38 crc kubenswrapper[4715]: I1125 12:26:38.310775 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8167972f-8b24-4c5d-ba63-fe81885b0e8c","Type":"ContainerStarted","Data":"6a9530b58a3a055c6b52f4803335174bb3cfc84bb1f2524a770bc5b19153e719"} Nov 25 12:26:38 crc kubenswrapper[4715]: I1125 12:26:38.703672 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7811654a-cd03-48f4-bb9d-1bd6cbf2094c" path="/var/lib/kubelet/pods/7811654a-cd03-48f4-bb9d-1bd6cbf2094c/volumes" Nov 25 12:26:39 crc kubenswrapper[4715]: I1125 12:26:39.324439 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8167972f-8b24-4c5d-ba63-fe81885b0e8c","Type":"ContainerStarted","Data":"4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41"} Nov 25 12:26:39 crc kubenswrapper[4715]: I1125 12:26:39.352106 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.352086499 podStartE2EDuration="2.352086499s" podCreationTimestamp="2025-11-25 12:26:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:26:39.34564984 +0000 UTC m=+1049.853152881" watchObservedRunningTime="2025-11-25 12:26:39.352086499 +0000 UTC m=+1049.859589520" Nov 25 12:26:47 crc kubenswrapper[4715]: I1125 12:26:47.687876 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:47 crc kubenswrapper[4715]: I1125 12:26:47.688686 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:47 crc kubenswrapper[4715]: I1125 12:26:47.725205 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:47 crc kubenswrapper[4715]: I1125 12:26:47.741656 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:47 crc kubenswrapper[4715]: I1125 12:26:47.944384 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:26:47 crc kubenswrapper[4715]: I1125 12:26:47.944478 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:26:47 crc kubenswrapper[4715]: I1125 12:26:47.944559 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:26:47 crc kubenswrapper[4715]: I1125 12:26:47.945658 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"244d5332aa55fb80cd40f8e16728cd8f4bf2022a5bf11e13456cf894fe4243d1"} pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:26:47 crc kubenswrapper[4715]: I1125 12:26:47.945774 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" containerID="cri-o://244d5332aa55fb80cd40f8e16728cd8f4bf2022a5bf11e13456cf894fe4243d1" gracePeriod=600 Nov 25 12:26:48 crc kubenswrapper[4715]: I1125 12:26:48.404862 4715 generic.go:334] "Generic (PLEG): container finished" podID="51676837-adc4-4424-b527-920a6528b6a2" containerID="244d5332aa55fb80cd40f8e16728cd8f4bf2022a5bf11e13456cf894fe4243d1" exitCode=0 Nov 25 12:26:48 crc kubenswrapper[4715]: I1125 12:26:48.404898 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerDied","Data":"244d5332aa55fb80cd40f8e16728cd8f4bf2022a5bf11e13456cf894fe4243d1"} Nov 25 12:26:48 crc kubenswrapper[4715]: I1125 12:26:48.405225 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"ad3f8b219f61d21b837d289c8ec7099946d5ca2ed30ff86a4db4d5d526737939"} Nov 25 12:26:48 crc kubenswrapper[4715]: I1125 12:26:48.405249 4715 scope.go:117] "RemoveContainer" containerID="23a04d777596661d2024d01da087796f35f637e5c71480856ee076f6f5b38b7a" Nov 25 12:26:48 crc kubenswrapper[4715]: I1125 12:26:48.405675 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:48 crc kubenswrapper[4715]: I1125 12:26:48.405704 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:50 crc kubenswrapper[4715]: I1125 12:26:50.378924 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:26:50 crc kubenswrapper[4715]: I1125 12:26:50.385531 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.116740 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-7xvp9"] Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.131778 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-7xvp9"] Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.220978 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-ckzgh"] Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.222500 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.225907 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.226166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.230276 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-ckzgh"] Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.349503 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-combined-ca-bundle\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.349581 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-db-sync-config-data\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.349597 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-config-data\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.349708 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx85b\" (UniqueName: \"kubernetes.io/projected/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-kube-api-access-tx85b\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.451052 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-db-sync-config-data\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.451105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-config-data\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.451252 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx85b\" (UniqueName: \"kubernetes.io/projected/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-kube-api-access-tx85b\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.451288 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-combined-ca-bundle\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.458388 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-combined-ca-bundle\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.462384 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-db-sync-config-data\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.463996 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-config-data\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.482658 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx85b\" (UniqueName: \"kubernetes.io/projected/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-kube-api-access-tx85b\") pod \"glance-db-sync-ckzgh\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.541731 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.707784 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31de582a-c3fe-412a-a6a6-fde502f793b1" path="/var/lib/kubelet/pods/31de582a-c3fe-412a-a6a6-fde502f793b1/volumes" Nov 25 12:27:00 crc kubenswrapper[4715]: I1125 12:27:00.981501 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-ckzgh"] Nov 25 12:27:01 crc kubenswrapper[4715]: I1125 12:27:01.537159 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-ckzgh" event={"ID":"97a2aa97-f8a8-44b4-b7cb-3339dfd26383","Type":"ContainerStarted","Data":"78dac332f9b3a5cdc999e2d58c6bffe4e91b09c113a258d3cd1aa176bb7da2ea"} Nov 25 12:27:01 crc kubenswrapper[4715]: I1125 12:27:01.538136 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-ckzgh" event={"ID":"97a2aa97-f8a8-44b4-b7cb-3339dfd26383","Type":"ContainerStarted","Data":"7ed92f94dc3c92680adb43421b7839789f8d132b7425616ecd50e5d3b1c24f22"} Nov 25 12:27:01 crc kubenswrapper[4715]: I1125 12:27:01.555805 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-ckzgh" podStartSLOduration=1.555782046 podStartE2EDuration="1.555782046s" podCreationTimestamp="2025-11-25 12:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:27:01.549916954 +0000 UTC m=+1072.057419975" watchObservedRunningTime="2025-11-25 12:27:01.555782046 +0000 UTC m=+1072.063285067" Nov 25 12:27:04 crc kubenswrapper[4715]: I1125 12:27:04.564853 4715 generic.go:334] "Generic (PLEG): container finished" podID="97a2aa97-f8a8-44b4-b7cb-3339dfd26383" containerID="78dac332f9b3a5cdc999e2d58c6bffe4e91b09c113a258d3cd1aa176bb7da2ea" exitCode=0 Nov 25 12:27:04 crc kubenswrapper[4715]: I1125 12:27:04.564942 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-ckzgh" event={"ID":"97a2aa97-f8a8-44b4-b7cb-3339dfd26383","Type":"ContainerDied","Data":"78dac332f9b3a5cdc999e2d58c6bffe4e91b09c113a258d3cd1aa176bb7da2ea"} Nov 25 12:27:05 crc kubenswrapper[4715]: I1125 12:27:05.877782 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:05 crc kubenswrapper[4715]: I1125 12:27:05.950167 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-db-sync-config-data\") pod \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " Nov 25 12:27:05 crc kubenswrapper[4715]: I1125 12:27:05.950260 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-config-data\") pod \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " Nov 25 12:27:05 crc kubenswrapper[4715]: I1125 12:27:05.950297 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx85b\" (UniqueName: \"kubernetes.io/projected/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-kube-api-access-tx85b\") pod \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " Nov 25 12:27:05 crc kubenswrapper[4715]: I1125 12:27:05.950380 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-combined-ca-bundle\") pod \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\" (UID: \"97a2aa97-f8a8-44b4-b7cb-3339dfd26383\") " Nov 25 12:27:05 crc kubenswrapper[4715]: I1125 12:27:05.956479 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-kube-api-access-tx85b" (OuterVolumeSpecName: "kube-api-access-tx85b") pod "97a2aa97-f8a8-44b4-b7cb-3339dfd26383" (UID: "97a2aa97-f8a8-44b4-b7cb-3339dfd26383"). InnerVolumeSpecName "kube-api-access-tx85b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:05 crc kubenswrapper[4715]: I1125 12:27:05.956523 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "97a2aa97-f8a8-44b4-b7cb-3339dfd26383" (UID: "97a2aa97-f8a8-44b4-b7cb-3339dfd26383"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:05 crc kubenswrapper[4715]: I1125 12:27:05.981374 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97a2aa97-f8a8-44b4-b7cb-3339dfd26383" (UID: "97a2aa97-f8a8-44b4-b7cb-3339dfd26383"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:05 crc kubenswrapper[4715]: I1125 12:27:05.991791 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-config-data" (OuterVolumeSpecName: "config-data") pod "97a2aa97-f8a8-44b4-b7cb-3339dfd26383" (UID: "97a2aa97-f8a8-44b4-b7cb-3339dfd26383"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.052720 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.052768 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.052784 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.052852 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx85b\" (UniqueName: \"kubernetes.io/projected/97a2aa97-f8a8-44b4-b7cb-3339dfd26383-kube-api-access-tx85b\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.588865 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-ckzgh" event={"ID":"97a2aa97-f8a8-44b4-b7cb-3339dfd26383","Type":"ContainerDied","Data":"7ed92f94dc3c92680adb43421b7839789f8d132b7425616ecd50e5d3b1c24f22"} Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.588905 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ed92f94dc3c92680adb43421b7839789f8d132b7425616ecd50e5d3b1c24f22" Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.588949 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-ckzgh" Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.702716 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.702982 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.703197 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-log" containerID="cri-o://82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336" gracePeriod=30 Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.703308 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-httpd" containerID="cri-o://4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41" gracePeriod=30 Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.703517 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerName="glance-log" containerID="cri-o://8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0" gracePeriod=30 Nov 25 12:27:06 crc kubenswrapper[4715]: I1125 12:27:06.703664 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerName="glance-httpd" containerID="cri-o://a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd" gracePeriod=30 Nov 25 12:27:07 crc kubenswrapper[4715]: I1125 12:27:07.598840 4715 generic.go:334] "Generic (PLEG): container finished" podID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerID="8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0" exitCode=143 Nov 25 12:27:07 crc kubenswrapper[4715]: I1125 12:27:07.598918 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"d4b54693-6831-4fb0-b631-7950e9a9234b","Type":"ContainerDied","Data":"8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0"} Nov 25 12:27:07 crc kubenswrapper[4715]: I1125 12:27:07.601097 4715 generic.go:334] "Generic (PLEG): container finished" podID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerID="82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336" exitCode=143 Nov 25 12:27:07 crc kubenswrapper[4715]: I1125 12:27:07.601157 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8167972f-8b24-4c5d-ba63-fe81885b0e8c","Type":"ContainerDied","Data":"82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336"} Nov 25 12:27:09 crc kubenswrapper[4715]: I1125 12:27:09.857371 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.102:9292/healthcheck\": read tcp 10.217.0.2:46294->10.217.0.102:9292: read: connection reset by peer" Nov 25 12:27:09 crc kubenswrapper[4715]: I1125 12:27:09.857599 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.102:9292/healthcheck\": read tcp 10.217.0.2:46304->10.217.0.102:9292: read: connection reset by peer" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.314757 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.322620 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421040 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-nvme\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421107 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-run\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421144 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421167 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-iscsi\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421204 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-sys\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421205 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421230 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-lib-modules\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421281 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-logs\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421309 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-lib-modules\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421330 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-dev\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421363 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjc56\" (UniqueName: \"kubernetes.io/projected/8167972f-8b24-4c5d-ba63-fe81885b0e8c-kube-api-access-zjc56\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421385 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421412 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvghk\" (UniqueName: \"kubernetes.io/projected/d4b54693-6831-4fb0-b631-7950e9a9234b-kube-api-access-lvghk\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421432 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-httpd-run\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421484 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-logs\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421511 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-var-locks-brick\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421529 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-var-locks-brick\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421543 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-run" (OuterVolumeSpecName: "run") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421554 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-scripts\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421582 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421606 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-dev\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421627 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-iscsi\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421664 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421693 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-config-data\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421717 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-httpd-run\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421739 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-sys\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421757 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-run\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421826 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-config-data\") pod \"d4b54693-6831-4fb0-b631-7950e9a9234b\" (UID: \"d4b54693-6831-4fb0-b631-7950e9a9234b\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421869 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-scripts\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421889 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-nvme\") pod \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\" (UID: \"8167972f-8b24-4c5d-ba63-fe81885b0e8c\") " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.421947 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422026 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-sys" (OuterVolumeSpecName: "sys") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422052 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422265 4715 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422283 4715 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422280 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-dev" (OuterVolumeSpecName: "dev") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422295 4715 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422343 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422372 4715 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-sys\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422402 4715 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422492 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-logs" (OuterVolumeSpecName: "logs") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422540 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422565 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-dev" (OuterVolumeSpecName: "dev") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422606 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422650 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-run" (OuterVolumeSpecName: "run") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422878 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.422915 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.423856 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.423893 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-sys" (OuterVolumeSpecName: "sys") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.424213 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-logs" (OuterVolumeSpecName: "logs") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.424580 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.427291 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-scripts" (OuterVolumeSpecName: "scripts") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.427981 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.427987 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8167972f-8b24-4c5d-ba63-fe81885b0e8c-kube-api-access-zjc56" (OuterVolumeSpecName: "kube-api-access-zjc56") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "kube-api-access-zjc56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.428170 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.428397 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.428558 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4b54693-6831-4fb0-b631-7950e9a9234b-kube-api-access-lvghk" (OuterVolumeSpecName: "kube-api-access-lvghk") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "kube-api-access-lvghk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.429372 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-scripts" (OuterVolumeSpecName: "scripts") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.439513 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.460434 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-config-data" (OuterVolumeSpecName: "config-data") pod "8167972f-8b24-4c5d-ba63-fe81885b0e8c" (UID: "8167972f-8b24-4c5d-ba63-fe81885b0e8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.463263 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-config-data" (OuterVolumeSpecName: "config-data") pod "d4b54693-6831-4fb0-b631-7950e9a9234b" (UID: "d4b54693-6831-4fb0-b631-7950e9a9234b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523156 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523225 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-logs\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523236 4715 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523245 4715 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-dev\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523256 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjc56\" (UniqueName: \"kubernetes.io/projected/8167972f-8b24-4c5d-ba63-fe81885b0e8c-kube-api-access-zjc56\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523273 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523282 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvghk\" (UniqueName: \"kubernetes.io/projected/d4b54693-6831-4fb0-b631-7950e9a9234b-kube-api-access-lvghk\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523291 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8167972f-8b24-4c5d-ba63-fe81885b0e8c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523300 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-logs\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523308 4715 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523317 4715 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523325 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523338 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523346 4715 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-dev\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523355 4715 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523367 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523375 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523383 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4b54693-6831-4fb0-b631-7950e9a9234b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523391 4715 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4b54693-6831-4fb0-b631-7950e9a9234b-sys\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523398 4715 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523406 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4b54693-6831-4fb0-b631-7950e9a9234b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523413 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8167972f-8b24-4c5d-ba63-fe81885b0e8c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.523420 4715 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8167972f-8b24-4c5d-ba63-fe81885b0e8c-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.535995 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.536809 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.536817 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.538388 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.624556 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.624589 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.624598 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.624608 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.628158 4715 generic.go:334] "Generic (PLEG): container finished" podID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerID="4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41" exitCode=0 Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.628254 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8167972f-8b24-4c5d-ba63-fe81885b0e8c","Type":"ContainerDied","Data":"4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41"} Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.628282 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.628312 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"8167972f-8b24-4c5d-ba63-fe81885b0e8c","Type":"ContainerDied","Data":"6a9530b58a3a055c6b52f4803335174bb3cfc84bb1f2524a770bc5b19153e719"} Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.628362 4715 scope.go:117] "RemoveContainer" containerID="4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.631109 4715 generic.go:334] "Generic (PLEG): container finished" podID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerID="a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd" exitCode=0 Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.631157 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.631176 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"d4b54693-6831-4fb0-b631-7950e9a9234b","Type":"ContainerDied","Data":"a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd"} Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.631638 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"d4b54693-6831-4fb0-b631-7950e9a9234b","Type":"ContainerDied","Data":"d43d83f1f58e59461b4fa33ac1ce2a190f98bb891b109ce967d0b231a050b79c"} Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.650571 4715 scope.go:117] "RemoveContainer" containerID="82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.675849 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.684576 4715 scope.go:117] "RemoveContainer" containerID="4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41" Nov 25 12:27:10 crc kubenswrapper[4715]: E1125 12:27:10.685178 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41\": container with ID starting with 4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41 not found: ID does not exist" containerID="4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.685269 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41"} err="failed to get container status \"4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41\": rpc error: code = NotFound desc = could not find container \"4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41\": container with ID starting with 4a7f18dacb48baab7295684860f4d1a8d8c93c9b50f5eea9c9151c26ddfa0c41 not found: ID does not exist" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.685315 4715 scope.go:117] "RemoveContainer" containerID="82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336" Nov 25 12:27:10 crc kubenswrapper[4715]: E1125 12:27:10.686553 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336\": container with ID starting with 82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336 not found: ID does not exist" containerID="82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.686603 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336"} err="failed to get container status \"82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336\": rpc error: code = NotFound desc = could not find container \"82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336\": container with ID starting with 82a207deba1e018608197143c4887343e78c8a28432b990a0463b5e01f9ce336 not found: ID does not exist" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.686630 4715 scope.go:117] "RemoveContainer" containerID="a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.713637 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.715836 4715 scope.go:117] "RemoveContainer" containerID="8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.715961 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.726396 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.732636 4715 scope.go:117] "RemoveContainer" containerID="a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd" Nov 25 12:27:10 crc kubenswrapper[4715]: E1125 12:27:10.733323 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd\": container with ID starting with a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd not found: ID does not exist" containerID="a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.733377 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd"} err="failed to get container status \"a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd\": rpc error: code = NotFound desc = could not find container \"a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd\": container with ID starting with a2f13dfbaeb4a1efebeea41f2d33e4eb995b2dd205e83a2e98a32c4a966be3dd not found: ID does not exist" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.733414 4715 scope.go:117] "RemoveContainer" containerID="8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0" Nov 25 12:27:10 crc kubenswrapper[4715]: E1125 12:27:10.733950 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0\": container with ID starting with 8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0 not found: ID does not exist" containerID="8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0" Nov 25 12:27:10 crc kubenswrapper[4715]: I1125 12:27:10.733975 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0"} err="failed to get container status \"8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0\": rpc error: code = NotFound desc = could not find container \"8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0\": container with ID starting with 8dc70e4738ea69068f7ea50cfbbc0aadd74490d212ce02b157a06d3f1c905bf0 not found: ID does not exist" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.970094 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:27:11 crc kubenswrapper[4715]: E1125 12:27:11.970793 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-httpd" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.970812 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-httpd" Nov 25 12:27:11 crc kubenswrapper[4715]: E1125 12:27:11.970834 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-log" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.970844 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-log" Nov 25 12:27:11 crc kubenswrapper[4715]: E1125 12:27:11.970864 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerName="glance-log" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.970874 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerName="glance-log" Nov 25 12:27:11 crc kubenswrapper[4715]: E1125 12:27:11.970896 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerName="glance-httpd" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.970907 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerName="glance-httpd" Nov 25 12:27:11 crc kubenswrapper[4715]: E1125 12:27:11.970928 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a2aa97-f8a8-44b4-b7cb-3339dfd26383" containerName="glance-db-sync" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.970938 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a2aa97-f8a8-44b4-b7cb-3339dfd26383" containerName="glance-db-sync" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.971121 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a2aa97-f8a8-44b4-b7cb-3339dfd26383" containerName="glance-db-sync" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.971140 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerName="glance-log" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.971155 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-httpd" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.971177 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4b54693-6831-4fb0-b631-7950e9a9234b" containerName="glance-httpd" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.971212 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" containerName="glance-log" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.972326 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.974166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.974785 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.975105 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.982058 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.984674 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-7gbkp" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.984959 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 25 12:27:11 crc kubenswrapper[4715]: I1125 12:27:11.996628 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.042660 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-config-data\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.042920 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqc29\" (UniqueName: \"kubernetes.io/projected/3a973e2e-0d25-4e54-92fb-80646d419739-kube-api-access-xqc29\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.043000 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-logs\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.043103 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.043244 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-scripts\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.043352 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.043448 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-httpd-run\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.043537 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.043613 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-config-data\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145542 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqc29\" (UniqueName: \"kubernetes.io/projected/3a973e2e-0d25-4e54-92fb-80646d419739-kube-api-access-xqc29\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145568 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-logs\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145589 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145627 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-scripts\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145646 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145676 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-httpd-run\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145713 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145730 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.145958 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.146666 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-httpd-run\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.146999 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-logs\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.154135 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.154233 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-scripts\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.154783 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.154923 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.157499 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-config-data\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.163327 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqc29\" (UniqueName: \"kubernetes.io/projected/3a973e2e-0d25-4e54-92fb-80646d419739-kube-api-access-xqc29\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.165844 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.292754 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.704675 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8167972f-8b24-4c5d-ba63-fe81885b0e8c" path="/var/lib/kubelet/pods/8167972f-8b24-4c5d-ba63-fe81885b0e8c/volumes" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.705979 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4b54693-6831-4fb0-b631-7950e9a9234b" path="/var/lib/kubelet/pods/d4b54693-6831-4fb0-b631-7950e9a9234b/volumes" Nov 25 12:27:12 crc kubenswrapper[4715]: I1125 12:27:12.763283 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:27:13 crc kubenswrapper[4715]: I1125 12:27:13.654179 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3a973e2e-0d25-4e54-92fb-80646d419739","Type":"ContainerStarted","Data":"0efe25f6b5193cd666ee0e78ff06bb1f23896d5beff658e4a459d6117df8d2ca"} Nov 25 12:27:13 crc kubenswrapper[4715]: I1125 12:27:13.654498 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3a973e2e-0d25-4e54-92fb-80646d419739","Type":"ContainerStarted","Data":"adb123e6826ef244dc9984c3dba00235a8ce560488e596b02134a80cca157593"} Nov 25 12:27:14 crc kubenswrapper[4715]: I1125 12:27:14.669908 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3a973e2e-0d25-4e54-92fb-80646d419739","Type":"ContainerStarted","Data":"bbc0ae1442d2275b48dcc3e9b65da7bdef05499cabf2547a7969573ae90d7f84"} Nov 25 12:27:14 crc kubenswrapper[4715]: I1125 12:27:14.691699 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.69167877 podStartE2EDuration="3.69167877s" podCreationTimestamp="2025-11-25 12:27:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:27:14.687657959 +0000 UTC m=+1085.195160980" watchObservedRunningTime="2025-11-25 12:27:14.69167877 +0000 UTC m=+1085.199181791" Nov 25 12:27:22 crc kubenswrapper[4715]: I1125 12:27:22.293123 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:22 crc kubenswrapper[4715]: I1125 12:27:22.293843 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:22 crc kubenswrapper[4715]: I1125 12:27:22.328325 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:22 crc kubenswrapper[4715]: I1125 12:27:22.328570 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:22 crc kubenswrapper[4715]: I1125 12:27:22.734784 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:22 crc kubenswrapper[4715]: I1125 12:27:22.734833 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:24 crc kubenswrapper[4715]: I1125 12:27:24.667875 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:24 crc kubenswrapper[4715]: I1125 12:27:24.684279 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.326053 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-ckzgh"] Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.333170 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-ckzgh"] Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.384596 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance63a3-account-delete-c2pcr"] Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.399112 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance63a3-account-delete-c2pcr"] Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.399230 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.415206 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.565274 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6hjw\" (UniqueName: \"kubernetes.io/projected/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-kube-api-access-b6hjw\") pod \"glance63a3-account-delete-c2pcr\" (UID: \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\") " pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.565329 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-operator-scripts\") pod \"glance63a3-account-delete-c2pcr\" (UID: \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\") " pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.666698 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6hjw\" (UniqueName: \"kubernetes.io/projected/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-kube-api-access-b6hjw\") pod \"glance63a3-account-delete-c2pcr\" (UID: \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\") " pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.666749 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-operator-scripts\") pod \"glance63a3-account-delete-c2pcr\" (UID: \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\") " pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.667450 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-operator-scripts\") pod \"glance63a3-account-delete-c2pcr\" (UID: \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\") " pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.686023 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6hjw\" (UniqueName: \"kubernetes.io/projected/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-kube-api-access-b6hjw\") pod \"glance63a3-account-delete-c2pcr\" (UID: \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\") " pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.702525 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a2aa97-f8a8-44b4-b7cb-3339dfd26383" path="/var/lib/kubelet/pods/97a2aa97-f8a8-44b4-b7cb-3339dfd26383/volumes" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.716525 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.765449 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="3a973e2e-0d25-4e54-92fb-80646d419739" containerName="glance-log" containerID="cri-o://0efe25f6b5193cd666ee0e78ff06bb1f23896d5beff658e4a459d6117df8d2ca" gracePeriod=30 Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.766015 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="3a973e2e-0d25-4e54-92fb-80646d419739" containerName="glance-httpd" containerID="cri-o://bbc0ae1442d2275b48dcc3e9b65da7bdef05499cabf2547a7969573ae90d7f84" gracePeriod=30 Nov 25 12:27:26 crc kubenswrapper[4715]: I1125 12:27:26.974009 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance63a3-account-delete-c2pcr"] Nov 25 12:27:26 crc kubenswrapper[4715]: W1125 12:27:26.982326 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81d070f3_f51b_4fd9_9da9_65ef2b3ae5db.slice/crio-366aa678f8819544af467acf5e9ee00546db0b5d556627a90e7f02f399b85bd1 WatchSource:0}: Error finding container 366aa678f8819544af467acf5e9ee00546db0b5d556627a90e7f02f399b85bd1: Status 404 returned error can't find the container with id 366aa678f8819544af467acf5e9ee00546db0b5d556627a90e7f02f399b85bd1 Nov 25 12:27:27 crc kubenswrapper[4715]: I1125 12:27:27.788458 4715 generic.go:334] "Generic (PLEG): container finished" podID="3a973e2e-0d25-4e54-92fb-80646d419739" containerID="0efe25f6b5193cd666ee0e78ff06bb1f23896d5beff658e4a459d6117df8d2ca" exitCode=143 Nov 25 12:27:27 crc kubenswrapper[4715]: I1125 12:27:27.788927 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3a973e2e-0d25-4e54-92fb-80646d419739","Type":"ContainerDied","Data":"0efe25f6b5193cd666ee0e78ff06bb1f23896d5beff658e4a459d6117df8d2ca"} Nov 25 12:27:27 crc kubenswrapper[4715]: I1125 12:27:27.794089 4715 generic.go:334] "Generic (PLEG): container finished" podID="81d070f3-f51b-4fd9-9da9-65ef2b3ae5db" containerID="ac107488d1aed9b03ae67615decb7c97a28dd3e761da5b2c0c06b442c8609752" exitCode=0 Nov 25 12:27:27 crc kubenswrapper[4715]: I1125 12:27:27.794142 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" event={"ID":"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db","Type":"ContainerDied","Data":"ac107488d1aed9b03ae67615decb7c97a28dd3e761da5b2c0c06b442c8609752"} Nov 25 12:27:27 crc kubenswrapper[4715]: I1125 12:27:27.794207 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" event={"ID":"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db","Type":"ContainerStarted","Data":"366aa678f8819544af467acf5e9ee00546db0b5d556627a90e7f02f399b85bd1"} Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.069521 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.207064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-operator-scripts\") pod \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\" (UID: \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\") " Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.207392 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6hjw\" (UniqueName: \"kubernetes.io/projected/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-kube-api-access-b6hjw\") pod \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\" (UID: \"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db\") " Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.208332 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81d070f3-f51b-4fd9-9da9-65ef2b3ae5db" (UID: "81d070f3-f51b-4fd9-9da9-65ef2b3ae5db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.212871 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-kube-api-access-b6hjw" (OuterVolumeSpecName: "kube-api-access-b6hjw") pod "81d070f3-f51b-4fd9-9da9-65ef2b3ae5db" (UID: "81d070f3-f51b-4fd9-9da9-65ef2b3ae5db"). InnerVolumeSpecName "kube-api-access-b6hjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.309608 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6hjw\" (UniqueName: \"kubernetes.io/projected/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-kube-api-access-b6hjw\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.309663 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.814512 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" event={"ID":"81d070f3-f51b-4fd9-9da9-65ef2b3ae5db","Type":"ContainerDied","Data":"366aa678f8819544af467acf5e9ee00546db0b5d556627a90e7f02f399b85bd1"} Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.814550 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="366aa678f8819544af467acf5e9ee00546db0b5d556627a90e7f02f399b85bd1" Nov 25 12:27:29 crc kubenswrapper[4715]: I1125 12:27:29.814628 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance63a3-account-delete-c2pcr" Nov 25 12:27:30 crc kubenswrapper[4715]: I1125 12:27:30.826083 4715 generic.go:334] "Generic (PLEG): container finished" podID="3a973e2e-0d25-4e54-92fb-80646d419739" containerID="bbc0ae1442d2275b48dcc3e9b65da7bdef05499cabf2547a7969573ae90d7f84" exitCode=0 Nov 25 12:27:30 crc kubenswrapper[4715]: I1125 12:27:30.826152 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3a973e2e-0d25-4e54-92fb-80646d419739","Type":"ContainerDied","Data":"bbc0ae1442d2275b48dcc3e9b65da7bdef05499cabf2547a7969573ae90d7f84"} Nov 25 12:27:30 crc kubenswrapper[4715]: I1125 12:27:30.826618 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"3a973e2e-0d25-4e54-92fb-80646d419739","Type":"ContainerDied","Data":"adb123e6826ef244dc9984c3dba00235a8ce560488e596b02134a80cca157593"} Nov 25 12:27:30 crc kubenswrapper[4715]: I1125 12:27:30.826632 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adb123e6826ef244dc9984c3dba00235a8ce560488e596b02134a80cca157593" Nov 25 12:27:30 crc kubenswrapper[4715]: I1125 12:27:30.834031 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.032486 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-combined-ca-bundle\") pod \"3a973e2e-0d25-4e54-92fb-80646d419739\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.032546 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqc29\" (UniqueName: \"kubernetes.io/projected/3a973e2e-0d25-4e54-92fb-80646d419739-kube-api-access-xqc29\") pod \"3a973e2e-0d25-4e54-92fb-80646d419739\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.032567 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-httpd-run\") pod \"3a973e2e-0d25-4e54-92fb-80646d419739\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.032592 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-internal-tls-certs\") pod \"3a973e2e-0d25-4e54-92fb-80646d419739\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.032651 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-config-data\") pod \"3a973e2e-0d25-4e54-92fb-80646d419739\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.032689 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-scripts\") pod \"3a973e2e-0d25-4e54-92fb-80646d419739\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.032913 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"3a973e2e-0d25-4e54-92fb-80646d419739\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.032973 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-public-tls-certs\") pod \"3a973e2e-0d25-4e54-92fb-80646d419739\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.033036 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-logs\") pod \"3a973e2e-0d25-4e54-92fb-80646d419739\" (UID: \"3a973e2e-0d25-4e54-92fb-80646d419739\") " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.033670 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3a973e2e-0d25-4e54-92fb-80646d419739" (UID: "3a973e2e-0d25-4e54-92fb-80646d419739"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.033830 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-logs" (OuterVolumeSpecName: "logs") pod "3a973e2e-0d25-4e54-92fb-80646d419739" (UID: "3a973e2e-0d25-4e54-92fb-80646d419739"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.039245 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a973e2e-0d25-4e54-92fb-80646d419739-kube-api-access-xqc29" (OuterVolumeSpecName: "kube-api-access-xqc29") pod "3a973e2e-0d25-4e54-92fb-80646d419739" (UID: "3a973e2e-0d25-4e54-92fb-80646d419739"). InnerVolumeSpecName "kube-api-access-xqc29". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.039559 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-scripts" (OuterVolumeSpecName: "scripts") pod "3a973e2e-0d25-4e54-92fb-80646d419739" (UID: "3a973e2e-0d25-4e54-92fb-80646d419739"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.045040 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "3a973e2e-0d25-4e54-92fb-80646d419739" (UID: "3a973e2e-0d25-4e54-92fb-80646d419739"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.058555 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a973e2e-0d25-4e54-92fb-80646d419739" (UID: "3a973e2e-0d25-4e54-92fb-80646d419739"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.073241 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3a973e2e-0d25-4e54-92fb-80646d419739" (UID: "3a973e2e-0d25-4e54-92fb-80646d419739"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.076311 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3a973e2e-0d25-4e54-92fb-80646d419739" (UID: "3a973e2e-0d25-4e54-92fb-80646d419739"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.078132 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-config-data" (OuterVolumeSpecName: "config-data") pod "3a973e2e-0d25-4e54-92fb-80646d419739" (UID: "3a973e2e-0d25-4e54-92fb-80646d419739"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.134421 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.134486 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.134497 4715 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.134509 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-logs\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.134518 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.134529 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqc29\" (UniqueName: \"kubernetes.io/projected/3a973e2e-0d25-4e54-92fb-80646d419739-kube-api-access-xqc29\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.134538 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3a973e2e-0d25-4e54-92fb-80646d419739-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.134558 4715 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.134567 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a973e2e-0d25-4e54-92fb-80646d419739-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.148273 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.236563 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.427335 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-czw8b"] Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.433261 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-czw8b"] Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.443029 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance63a3-account-delete-c2pcr"] Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.448396 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-63a3-account-create-update-lfs2c"] Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.453648 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-63a3-account-create-update-lfs2c"] Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.458463 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance63a3-account-delete-c2pcr"] Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.836991 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.883785 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:27:31 crc kubenswrapper[4715]: I1125 12:27:31.888466 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.500726 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-4qnbn"] Nov 25 12:27:32 crc kubenswrapper[4715]: E1125 12:27:32.501101 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a973e2e-0d25-4e54-92fb-80646d419739" containerName="glance-httpd" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.501119 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a973e2e-0d25-4e54-92fb-80646d419739" containerName="glance-httpd" Nov 25 12:27:32 crc kubenswrapper[4715]: E1125 12:27:32.501132 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a973e2e-0d25-4e54-92fb-80646d419739" containerName="glance-log" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.501139 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a973e2e-0d25-4e54-92fb-80646d419739" containerName="glance-log" Nov 25 12:27:32 crc kubenswrapper[4715]: E1125 12:27:32.501155 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81d070f3-f51b-4fd9-9da9-65ef2b3ae5db" containerName="mariadb-account-delete" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.501164 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="81d070f3-f51b-4fd9-9da9-65ef2b3ae5db" containerName="mariadb-account-delete" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.501336 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a973e2e-0d25-4e54-92fb-80646d419739" containerName="glance-httpd" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.501357 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="81d070f3-f51b-4fd9-9da9-65ef2b3ae5db" containerName="mariadb-account-delete" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.501377 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a973e2e-0d25-4e54-92fb-80646d419739" containerName="glance-log" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.501966 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.509086 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd"] Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.510554 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.514275 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.518062 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd"] Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.523848 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-4qnbn"] Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.655401 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a847f70-4c70-4d94-8336-6c81d3697ac7-operator-scripts\") pod \"glance-db-create-4qnbn\" (UID: \"2a847f70-4c70-4d94-8336-6c81d3697ac7\") " pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.655555 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78827\" (UniqueName: \"kubernetes.io/projected/2a847f70-4c70-4d94-8336-6c81d3697ac7-kube-api-access-78827\") pod \"glance-db-create-4qnbn\" (UID: \"2a847f70-4c70-4d94-8336-6c81d3697ac7\") " pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.655629 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k8cz\" (UniqueName: \"kubernetes.io/projected/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-kube-api-access-9k8cz\") pod \"glance-bcbc-account-create-update-xhzzd\" (UID: \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\") " pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.655740 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-operator-scripts\") pod \"glance-bcbc-account-create-update-xhzzd\" (UID: \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\") " pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.704286 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a973e2e-0d25-4e54-92fb-80646d419739" path="/var/lib/kubelet/pods/3a973e2e-0d25-4e54-92fb-80646d419739/volumes" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.705040 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="676324c4-624e-44b5-adf0-f22c0e724765" path="/var/lib/kubelet/pods/676324c4-624e-44b5-adf0-f22c0e724765/volumes" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.705881 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81d070f3-f51b-4fd9-9da9-65ef2b3ae5db" path="/var/lib/kubelet/pods/81d070f3-f51b-4fd9-9da9-65ef2b3ae5db/volumes" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.706940 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a82194-e1ef-4c16-ba40-fcbc435a029a" path="/var/lib/kubelet/pods/b9a82194-e1ef-4c16-ba40-fcbc435a029a/volumes" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.757883 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a847f70-4c70-4d94-8336-6c81d3697ac7-operator-scripts\") pod \"glance-db-create-4qnbn\" (UID: \"2a847f70-4c70-4d94-8336-6c81d3697ac7\") " pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.758340 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78827\" (UniqueName: \"kubernetes.io/projected/2a847f70-4c70-4d94-8336-6c81d3697ac7-kube-api-access-78827\") pod \"glance-db-create-4qnbn\" (UID: \"2a847f70-4c70-4d94-8336-6c81d3697ac7\") " pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.758483 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k8cz\" (UniqueName: \"kubernetes.io/projected/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-kube-api-access-9k8cz\") pod \"glance-bcbc-account-create-update-xhzzd\" (UID: \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\") " pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.758539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-operator-scripts\") pod \"glance-bcbc-account-create-update-xhzzd\" (UID: \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\") " pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.758811 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a847f70-4c70-4d94-8336-6c81d3697ac7-operator-scripts\") pod \"glance-db-create-4qnbn\" (UID: \"2a847f70-4c70-4d94-8336-6c81d3697ac7\") " pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.759391 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-operator-scripts\") pod \"glance-bcbc-account-create-update-xhzzd\" (UID: \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\") " pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.776883 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78827\" (UniqueName: \"kubernetes.io/projected/2a847f70-4c70-4d94-8336-6c81d3697ac7-kube-api-access-78827\") pod \"glance-db-create-4qnbn\" (UID: \"2a847f70-4c70-4d94-8336-6c81d3697ac7\") " pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.777729 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k8cz\" (UniqueName: \"kubernetes.io/projected/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-kube-api-access-9k8cz\") pod \"glance-bcbc-account-create-update-xhzzd\" (UID: \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\") " pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.821206 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:32 crc kubenswrapper[4715]: I1125 12:27:32.831825 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:33 crc kubenswrapper[4715]: I1125 12:27:33.264971 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd"] Nov 25 12:27:33 crc kubenswrapper[4715]: W1125 12:27:33.270955 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0aa4f353_6207_4a3c_9e0c_4d04cefdab19.slice/crio-903beedf7d7a234083cd9b9270e3b9be2cc969b1e3c59a7f66eef62dac4b3073 WatchSource:0}: Error finding container 903beedf7d7a234083cd9b9270e3b9be2cc969b1e3c59a7f66eef62dac4b3073: Status 404 returned error can't find the container with id 903beedf7d7a234083cd9b9270e3b9be2cc969b1e3c59a7f66eef62dac4b3073 Nov 25 12:27:33 crc kubenswrapper[4715]: I1125 12:27:33.298283 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-4qnbn"] Nov 25 12:27:33 crc kubenswrapper[4715]: W1125 12:27:33.304821 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a847f70_4c70_4d94_8336_6c81d3697ac7.slice/crio-216f1c14610819c91d4aeaf89bc3b2fb782afb3cb1950ae169906bcea6b2f5aa WatchSource:0}: Error finding container 216f1c14610819c91d4aeaf89bc3b2fb782afb3cb1950ae169906bcea6b2f5aa: Status 404 returned error can't find the container with id 216f1c14610819c91d4aeaf89bc3b2fb782afb3cb1950ae169906bcea6b2f5aa Nov 25 12:27:33 crc kubenswrapper[4715]: I1125 12:27:33.858795 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" event={"ID":"0aa4f353-6207-4a3c-9e0c-4d04cefdab19","Type":"ContainerStarted","Data":"fcda9cb1b3598597ada716347e9305ee89b80141884f071771dcf4703dfd173c"} Nov 25 12:27:33 crc kubenswrapper[4715]: I1125 12:27:33.859176 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" event={"ID":"0aa4f353-6207-4a3c-9e0c-4d04cefdab19","Type":"ContainerStarted","Data":"903beedf7d7a234083cd9b9270e3b9be2cc969b1e3c59a7f66eef62dac4b3073"} Nov 25 12:27:33 crc kubenswrapper[4715]: I1125 12:27:33.861670 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4qnbn" event={"ID":"2a847f70-4c70-4d94-8336-6c81d3697ac7","Type":"ContainerStarted","Data":"f76baf4c89d00d914f98431134e489645a68c5b51a98f3bd126da05052fa8407"} Nov 25 12:27:33 crc kubenswrapper[4715]: I1125 12:27:33.861738 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4qnbn" event={"ID":"2a847f70-4c70-4d94-8336-6c81d3697ac7","Type":"ContainerStarted","Data":"216f1c14610819c91d4aeaf89bc3b2fb782afb3cb1950ae169906bcea6b2f5aa"} Nov 25 12:27:33 crc kubenswrapper[4715]: I1125 12:27:33.880770 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" podStartSLOduration=1.880751686 podStartE2EDuration="1.880751686s" podCreationTimestamp="2025-11-25 12:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:27:33.873886646 +0000 UTC m=+1104.381389697" watchObservedRunningTime="2025-11-25 12:27:33.880751686 +0000 UTC m=+1104.388254707" Nov 25 12:27:33 crc kubenswrapper[4715]: I1125 12:27:33.896261 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-4qnbn" podStartSLOduration=1.896237264 podStartE2EDuration="1.896237264s" podCreationTimestamp="2025-11-25 12:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:27:33.891261266 +0000 UTC m=+1104.398764287" watchObservedRunningTime="2025-11-25 12:27:33.896237264 +0000 UTC m=+1104.403740285" Nov 25 12:27:34 crc kubenswrapper[4715]: I1125 12:27:34.870267 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a847f70-4c70-4d94-8336-6c81d3697ac7" containerID="f76baf4c89d00d914f98431134e489645a68c5b51a98f3bd126da05052fa8407" exitCode=0 Nov 25 12:27:34 crc kubenswrapper[4715]: I1125 12:27:34.870347 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4qnbn" event={"ID":"2a847f70-4c70-4d94-8336-6c81d3697ac7","Type":"ContainerDied","Data":"f76baf4c89d00d914f98431134e489645a68c5b51a98f3bd126da05052fa8407"} Nov 25 12:27:34 crc kubenswrapper[4715]: I1125 12:27:34.872243 4715 generic.go:334] "Generic (PLEG): container finished" podID="0aa4f353-6207-4a3c-9e0c-4d04cefdab19" containerID="fcda9cb1b3598597ada716347e9305ee89b80141884f071771dcf4703dfd173c" exitCode=0 Nov 25 12:27:34 crc kubenswrapper[4715]: I1125 12:27:34.872307 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" event={"ID":"0aa4f353-6207-4a3c-9e0c-4d04cefdab19","Type":"ContainerDied","Data":"fcda9cb1b3598597ada716347e9305ee89b80141884f071771dcf4703dfd173c"} Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.220742 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.298125 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.413459 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k8cz\" (UniqueName: \"kubernetes.io/projected/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-kube-api-access-9k8cz\") pod \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\" (UID: \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\") " Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.413535 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78827\" (UniqueName: \"kubernetes.io/projected/2a847f70-4c70-4d94-8336-6c81d3697ac7-kube-api-access-78827\") pod \"2a847f70-4c70-4d94-8336-6c81d3697ac7\" (UID: \"2a847f70-4c70-4d94-8336-6c81d3697ac7\") " Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.413590 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-operator-scripts\") pod \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\" (UID: \"0aa4f353-6207-4a3c-9e0c-4d04cefdab19\") " Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.413749 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a847f70-4c70-4d94-8336-6c81d3697ac7-operator-scripts\") pod \"2a847f70-4c70-4d94-8336-6c81d3697ac7\" (UID: \"2a847f70-4c70-4d94-8336-6c81d3697ac7\") " Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.414467 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0aa4f353-6207-4a3c-9e0c-4d04cefdab19" (UID: "0aa4f353-6207-4a3c-9e0c-4d04cefdab19"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.414501 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a847f70-4c70-4d94-8336-6c81d3697ac7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2a847f70-4c70-4d94-8336-6c81d3697ac7" (UID: "2a847f70-4c70-4d94-8336-6c81d3697ac7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.422413 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-kube-api-access-9k8cz" (OuterVolumeSpecName: "kube-api-access-9k8cz") pod "0aa4f353-6207-4a3c-9e0c-4d04cefdab19" (UID: "0aa4f353-6207-4a3c-9e0c-4d04cefdab19"). InnerVolumeSpecName "kube-api-access-9k8cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.441454 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a847f70-4c70-4d94-8336-6c81d3697ac7-kube-api-access-78827" (OuterVolumeSpecName: "kube-api-access-78827") pod "2a847f70-4c70-4d94-8336-6c81d3697ac7" (UID: "2a847f70-4c70-4d94-8336-6c81d3697ac7"). InnerVolumeSpecName "kube-api-access-78827". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.515511 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a847f70-4c70-4d94-8336-6c81d3697ac7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.515549 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k8cz\" (UniqueName: \"kubernetes.io/projected/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-kube-api-access-9k8cz\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.515561 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78827\" (UniqueName: \"kubernetes.io/projected/2a847f70-4c70-4d94-8336-6c81d3697ac7-kube-api-access-78827\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.515570 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa4f353-6207-4a3c-9e0c-4d04cefdab19-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.892748 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" event={"ID":"0aa4f353-6207-4a3c-9e0c-4d04cefdab19","Type":"ContainerDied","Data":"903beedf7d7a234083cd9b9270e3b9be2cc969b1e3c59a7f66eef62dac4b3073"} Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.892780 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.892798 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="903beedf7d7a234083cd9b9270e3b9be2cc969b1e3c59a7f66eef62dac4b3073" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.895047 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4qnbn" event={"ID":"2a847f70-4c70-4d94-8336-6c81d3697ac7","Type":"ContainerDied","Data":"216f1c14610819c91d4aeaf89bc3b2fb782afb3cb1950ae169906bcea6b2f5aa"} Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.895075 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="216f1c14610819c91d4aeaf89bc3b2fb782afb3cb1950ae169906bcea6b2f5aa" Nov 25 12:27:36 crc kubenswrapper[4715]: I1125 12:27:36.895134 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4qnbn" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.701701 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-hxbr7"] Nov 25 12:27:37 crc kubenswrapper[4715]: E1125 12:27:37.702313 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a847f70-4c70-4d94-8336-6c81d3697ac7" containerName="mariadb-database-create" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.702329 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a847f70-4c70-4d94-8336-6c81d3697ac7" containerName="mariadb-database-create" Nov 25 12:27:37 crc kubenswrapper[4715]: E1125 12:27:37.702345 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa4f353-6207-4a3c-9e0c-4d04cefdab19" containerName="mariadb-account-create-update" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.702355 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa4f353-6207-4a3c-9e0c-4d04cefdab19" containerName="mariadb-account-create-update" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.702552 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a847f70-4c70-4d94-8336-6c81d3697ac7" containerName="mariadb-database-create" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.702583 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa4f353-6207-4a3c-9e0c-4d04cefdab19" containerName="mariadb-account-create-update" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.703172 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.705388 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.705539 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-f4vr5" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.717333 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hxbr7"] Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.731509 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpt7j\" (UniqueName: \"kubernetes.io/projected/77bf90c9-c3bc-4691-811d-975a2717e876-kube-api-access-dpt7j\") pod \"glance-db-sync-hxbr7\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.731595 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-db-sync-config-data\") pod \"glance-db-sync-hxbr7\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.731696 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-config-data\") pod \"glance-db-sync-hxbr7\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.833180 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-config-data\") pod \"glance-db-sync-hxbr7\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.833267 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpt7j\" (UniqueName: \"kubernetes.io/projected/77bf90c9-c3bc-4691-811d-975a2717e876-kube-api-access-dpt7j\") pod \"glance-db-sync-hxbr7\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.833314 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-db-sync-config-data\") pod \"glance-db-sync-hxbr7\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.839360 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-db-sync-config-data\") pod \"glance-db-sync-hxbr7\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.846200 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-config-data\") pod \"glance-db-sync-hxbr7\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:37 crc kubenswrapper[4715]: I1125 12:27:37.848955 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpt7j\" (UniqueName: \"kubernetes.io/projected/77bf90c9-c3bc-4691-811d-975a2717e876-kube-api-access-dpt7j\") pod \"glance-db-sync-hxbr7\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:38 crc kubenswrapper[4715]: I1125 12:27:38.019249 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:38 crc kubenswrapper[4715]: I1125 12:27:38.331847 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hxbr7"] Nov 25 12:27:38 crc kubenswrapper[4715]: W1125 12:27:38.336040 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77bf90c9_c3bc_4691_811d_975a2717e876.slice/crio-1092d587d0ff37774b328c935211b52d60c70130db03ab0292f9aa1ab590bd76 WatchSource:0}: Error finding container 1092d587d0ff37774b328c935211b52d60c70130db03ab0292f9aa1ab590bd76: Status 404 returned error can't find the container with id 1092d587d0ff37774b328c935211b52d60c70130db03ab0292f9aa1ab590bd76 Nov 25 12:27:38 crc kubenswrapper[4715]: I1125 12:27:38.911557 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hxbr7" event={"ID":"77bf90c9-c3bc-4691-811d-975a2717e876","Type":"ContainerStarted","Data":"1f10f0aa130071966a58980829c521daba9639f4201110251d7c3e04ad0f0087"} Nov 25 12:27:38 crc kubenswrapper[4715]: I1125 12:27:38.911609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hxbr7" event={"ID":"77bf90c9-c3bc-4691-811d-975a2717e876","Type":"ContainerStarted","Data":"1092d587d0ff37774b328c935211b52d60c70130db03ab0292f9aa1ab590bd76"} Nov 25 12:27:38 crc kubenswrapper[4715]: I1125 12:27:38.930071 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-hxbr7" podStartSLOduration=1.930044955 podStartE2EDuration="1.930044955s" podCreationTimestamp="2025-11-25 12:27:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:27:38.929313485 +0000 UTC m=+1109.436816536" watchObservedRunningTime="2025-11-25 12:27:38.930044955 +0000 UTC m=+1109.437547986" Nov 25 12:27:42 crc kubenswrapper[4715]: I1125 12:27:42.942256 4715 generic.go:334] "Generic (PLEG): container finished" podID="77bf90c9-c3bc-4691-811d-975a2717e876" containerID="1f10f0aa130071966a58980829c521daba9639f4201110251d7c3e04ad0f0087" exitCode=0 Nov 25 12:27:42 crc kubenswrapper[4715]: I1125 12:27:42.942368 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hxbr7" event={"ID":"77bf90c9-c3bc-4691-811d-975a2717e876","Type":"ContainerDied","Data":"1f10f0aa130071966a58980829c521daba9639f4201110251d7c3e04ad0f0087"} Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.238458 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.413168 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-config-data\") pod \"77bf90c9-c3bc-4691-811d-975a2717e876\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.413302 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-db-sync-config-data\") pod \"77bf90c9-c3bc-4691-811d-975a2717e876\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.413345 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpt7j\" (UniqueName: \"kubernetes.io/projected/77bf90c9-c3bc-4691-811d-975a2717e876-kube-api-access-dpt7j\") pod \"77bf90c9-c3bc-4691-811d-975a2717e876\" (UID: \"77bf90c9-c3bc-4691-811d-975a2717e876\") " Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.418225 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "77bf90c9-c3bc-4691-811d-975a2717e876" (UID: "77bf90c9-c3bc-4691-811d-975a2717e876"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.418638 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77bf90c9-c3bc-4691-811d-975a2717e876-kube-api-access-dpt7j" (OuterVolumeSpecName: "kube-api-access-dpt7j") pod "77bf90c9-c3bc-4691-811d-975a2717e876" (UID: "77bf90c9-c3bc-4691-811d-975a2717e876"). InnerVolumeSpecName "kube-api-access-dpt7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.452157 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-config-data" (OuterVolumeSpecName: "config-data") pod "77bf90c9-c3bc-4691-811d-975a2717e876" (UID: "77bf90c9-c3bc-4691-811d-975a2717e876"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.514510 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.514576 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/77bf90c9-c3bc-4691-811d-975a2717e876-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.514594 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpt7j\" (UniqueName: \"kubernetes.io/projected/77bf90c9-c3bc-4691-811d-975a2717e876-kube-api-access-dpt7j\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.963482 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hxbr7" event={"ID":"77bf90c9-c3bc-4691-811d-975a2717e876","Type":"ContainerDied","Data":"1092d587d0ff37774b328c935211b52d60c70130db03ab0292f9aa1ab590bd76"} Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.963549 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1092d587d0ff37774b328c935211b52d60c70130db03ab0292f9aa1ab590bd76" Nov 25 12:27:44 crc kubenswrapper[4715]: I1125 12:27:44.963657 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hxbr7" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.302918 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 12:27:46 crc kubenswrapper[4715]: E1125 12:27:46.303600 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77bf90c9-c3bc-4691-811d-975a2717e876" containerName="glance-db-sync" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.303617 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="77bf90c9-c3bc-4691-811d-975a2717e876" containerName="glance-db-sync" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.303780 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="77bf90c9-c3bc-4691-811d-975a2717e876" containerName="glance-db-sync" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.304982 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.307424 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.307675 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.308101 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-f4vr5" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.325711 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 12:27:46 crc kubenswrapper[4715]: E1125 12:27:46.326203 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data dev etc-iscsi etc-nvme glance glance-cache httpd-run kube-api-access-48hx5 lib-modules logs run scripts sys var-locks-brick], unattached volumes=[], failed to process volumes=[config-data dev etc-iscsi etc-nvme glance glance-cache httpd-run kube-api-access-48hx5 lib-modules logs run scripts sys var-locks-brick]: context canceled" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="da516516-c7f5-4e7e-b2d5-a3b540d31736" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.384589 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.386347 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.393691 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.396771 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.424781 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442268 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-logs\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442326 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-sys\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442399 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442433 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-dev\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442458 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442496 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-run\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442523 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48hx5\" (UniqueName: \"kubernetes.io/projected/da516516-c7f5-4e7e-b2d5-a3b540d31736-kube-api-access-48hx5\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442560 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-scripts\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442603 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442629 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442650 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442680 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-config-data\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.442766 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.543962 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544003 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m75h5\" (UniqueName: \"kubernetes.io/projected/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-kube-api-access-m75h5\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544032 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544063 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-dev\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544082 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544125 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-scripts\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544144 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-run\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544158 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48hx5\" (UniqueName: \"kubernetes.io/projected/da516516-c7f5-4e7e-b2d5-a3b540d31736-kube-api-access-48hx5\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544188 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544195 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-scripts\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544313 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-dev\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544356 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-run\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544385 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-dev\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544425 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544470 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-run\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544517 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544535 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-sys\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544560 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544575 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-config-data\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544618 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544633 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544656 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-config-data\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544677 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544696 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544720 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544752 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-logs\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544767 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544790 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-sys\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544818 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544900 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-logs\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544950 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.544909 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.545024 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.545130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-sys\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.545689 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-logs\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.545707 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.545758 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.550638 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-config-data\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.556726 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-scripts\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.568603 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48hx5\" (UniqueName: \"kubernetes.io/projected/da516516-c7f5-4e7e-b2d5-a3b540d31736-kube-api-access-48hx5\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.573818 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.575670 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646068 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-scripts\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646139 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-dev\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646167 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646192 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-run\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646244 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-sys\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646281 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-config-data\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646303 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646329 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646352 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646386 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646419 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646464 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-logs\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646495 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.646520 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m75h5\" (UniqueName: \"kubernetes.io/projected/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-kube-api-access-m75h5\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647189 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647394 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-dev\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647423 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-run\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647487 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647544 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647574 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647756 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647863 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647925 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647949 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-logs\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.647998 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-sys\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.664263 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-scripts\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.664680 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-config-data\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.680918 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.681687 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m75h5\" (UniqueName: \"kubernetes.io/projected/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-kube-api-access-m75h5\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.689014 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.703692 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.987134 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:46 crc kubenswrapper[4715]: I1125 12:27:46.997432 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154596 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-sys\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154674 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-var-locks-brick\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154727 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-config-data\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154753 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-lib-modules\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154761 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-sys" (OuterVolumeSpecName: "sys") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154794 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-scripts\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154821 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154844 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-dev\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154849 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154868 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-run\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154901 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-dev" (OuterVolumeSpecName: "dev") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154903 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-iscsi\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154891 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154928 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154931 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-run" (OuterVolumeSpecName: "run") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154950 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.154994 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48hx5\" (UniqueName: \"kubernetes.io/projected/da516516-c7f5-4e7e-b2d5-a3b540d31736-kube-api-access-48hx5\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155033 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-logs\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155098 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-nvme\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155144 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-httpd-run\") pod \"da516516-c7f5-4e7e-b2d5-a3b540d31736\" (UID: \"da516516-c7f5-4e7e-b2d5-a3b540d31736\") " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155463 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155520 4715 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155536 4715 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155550 4715 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-dev\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155546 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-logs" (OuterVolumeSpecName: "logs") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155562 4715 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155614 4715 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155631 4715 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-sys\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.155775 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.160128 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-config-data" (OuterVolumeSpecName: "config-data") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.160169 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-scripts" (OuterVolumeSpecName: "scripts") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.160169 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.160393 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.162328 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da516516-c7f5-4e7e-b2d5-a3b540d31736-kube-api-access-48hx5" (OuterVolumeSpecName: "kube-api-access-48hx5") pod "da516516-c7f5-4e7e-b2d5-a3b540d31736" (UID: "da516516-c7f5-4e7e-b2d5-a3b540d31736"). InnerVolumeSpecName "kube-api-access-48hx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.191519 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.239840 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.257386 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.257432 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.257446 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.257457 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48hx5\" (UniqueName: \"kubernetes.io/projected/da516516-c7f5-4e7e-b2d5-a3b540d31736-kube-api-access-48hx5\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.257466 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-logs\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.257475 4715 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/da516516-c7f5-4e7e-b2d5-a3b540d31736-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.257483 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da516516-c7f5-4e7e-b2d5-a3b540d31736-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.257492 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da516516-c7f5-4e7e-b2d5-a3b540d31736-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.269704 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.271607 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.359015 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.359052 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.997628 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.997657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37","Type":"ContainerStarted","Data":"6f18b54fcfb88137d23c27bccd1816fb2d373bff598f0e9ed83a62d49b2bd66c"} Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.998348 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37","Type":"ContainerStarted","Data":"00e3e7cb4096fb97a41463ef8618ce8d229124b5e060b72d25aaf4f072ed641a"} Nov 25 12:27:47 crc kubenswrapper[4715]: I1125 12:27:47.998365 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37","Type":"ContainerStarted","Data":"30824e67dc73a4b4127bbe682f06e76fae36975d81cef7dfa488895b41b2d6dd"} Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.085581 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.107281 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.126843 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.128369 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.131739 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.150113 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.280228 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.280817 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.280857 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.280881 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/744867e9-9122-4c3d-96a1-73156d822e67-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.280903 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-sys\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.280922 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.280978 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.281003 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/744867e9-9122-4c3d-96a1-73156d822e67-scripts\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.281083 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.281121 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/744867e9-9122-4c3d-96a1-73156d822e67-logs\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.281152 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/744867e9-9122-4c3d-96a1-73156d822e67-config-data\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.281177 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-dev\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.281202 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59jz7\" (UniqueName: \"kubernetes.io/projected/744867e9-9122-4c3d-96a1-73156d822e67-kube-api-access-59jz7\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.281268 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-run\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.382961 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.383674 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/744867e9-9122-4c3d-96a1-73156d822e67-scripts\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.383728 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.383768 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/744867e9-9122-4c3d-96a1-73156d822e67-logs\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.383812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/744867e9-9122-4c3d-96a1-73156d822e67-config-data\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.383839 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-dev\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.383869 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59jz7\" (UniqueName: \"kubernetes.io/projected/744867e9-9122-4c3d-96a1-73156d822e67-kube-api-access-59jz7\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.383936 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-run\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.383979 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.384028 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.384064 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.384097 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/744867e9-9122-4c3d-96a1-73156d822e67-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.384139 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-sys\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.384171 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.384312 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.383145 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.385187 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-run\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.385320 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.385429 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.385486 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-sys\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.385521 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.385653 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.385765 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/744867e9-9122-4c3d-96a1-73156d822e67-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.385961 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/744867e9-9122-4c3d-96a1-73156d822e67-logs\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.388752 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/744867e9-9122-4c3d-96a1-73156d822e67-dev\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.391060 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/744867e9-9122-4c3d-96a1-73156d822e67-scripts\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.391267 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/744867e9-9122-4c3d-96a1-73156d822e67-config-data\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.403999 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59jz7\" (UniqueName: \"kubernetes.io/projected/744867e9-9122-4c3d-96a1-73156d822e67-kube-api-access-59jz7\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.407362 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.414709 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"744867e9-9122-4c3d-96a1-73156d822e67\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:48 crc kubenswrapper[4715]: I1125 12:27:48.451474 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:48.702684 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da516516-c7f5-4e7e-b2d5-a3b540d31736" path="/var/lib/kubelet/pods/da516516-c7f5-4e7e-b2d5-a3b540d31736/volumes" Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:49.012593 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37","Type":"ContainerStarted","Data":"32c6a5c0dcde7b74f2b6d8e0dfc907a032d3fb0e7b760a6351d7387275bc12b5"} Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:49.012850 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-log" containerID="cri-o://00e3e7cb4096fb97a41463ef8618ce8d229124b5e060b72d25aaf4f072ed641a" gracePeriod=30 Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:49.013804 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-api" containerID="cri-o://32c6a5c0dcde7b74f2b6d8e0dfc907a032d3fb0e7b760a6351d7387275bc12b5" gracePeriod=30 Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:49.013948 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-httpd" containerID="cri-o://6f18b54fcfb88137d23c27bccd1816fb2d373bff598f0e9ed83a62d49b2bd66c" gracePeriod=30 Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:49.046956 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=4.046941527 podStartE2EDuration="4.046941527s" podCreationTimestamp="2025-11-25 12:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:27:49.046559957 +0000 UTC m=+1119.554062978" watchObservedRunningTime="2025-11-25 12:27:49.046941527 +0000 UTC m=+1119.554444548" Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:50.020197 4715 generic.go:334] "Generic (PLEG): container finished" podID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerID="32c6a5c0dcde7b74f2b6d8e0dfc907a032d3fb0e7b760a6351d7387275bc12b5" exitCode=143 Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:50.020491 4715 generic.go:334] "Generic (PLEG): container finished" podID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerID="6f18b54fcfb88137d23c27bccd1816fb2d373bff598f0e9ed83a62d49b2bd66c" exitCode=143 Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:50.020226 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37","Type":"ContainerDied","Data":"32c6a5c0dcde7b74f2b6d8e0dfc907a032d3fb0e7b760a6351d7387275bc12b5"} Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:50.020529 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37","Type":"ContainerDied","Data":"6f18b54fcfb88137d23c27bccd1816fb2d373bff598f0e9ed83a62d49b2bd66c"} Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:50.020544 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37","Type":"ContainerDied","Data":"00e3e7cb4096fb97a41463ef8618ce8d229124b5e060b72d25aaf4f072ed641a"} Nov 25 12:27:50 crc kubenswrapper[4715]: I1125 12:27:50.020500 4715 generic.go:334] "Generic (PLEG): container finished" podID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerID="00e3e7cb4096fb97a41463ef8618ce8d229124b5e060b72d25aaf4f072ed641a" exitCode=143 Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.117262 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.221102 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226053 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-run\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226106 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226131 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-iscsi\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226169 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226187 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-dev\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226198 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-run" (OuterVolumeSpecName: "run") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226219 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-sys\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226256 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-sys" (OuterVolumeSpecName: "sys") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226289 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226330 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m75h5\" (UniqueName: \"kubernetes.io/projected/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-kube-api-access-m75h5\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226381 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-nvme\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226405 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-scripts\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226423 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-var-locks-brick\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226455 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-lib-modules\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226533 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-logs\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226599 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-config-data\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226670 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-httpd-run\") pod \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\" (UID: \"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37\") " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.226859 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.227233 4715 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.227262 4715 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.227281 4715 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-sys\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.227296 4715 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.227335 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.227809 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-logs" (OuterVolumeSpecName: "logs") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.227867 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.228325 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-dev" (OuterVolumeSpecName: "dev") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.228993 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.231559 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.231651 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-kube-api-access-m75h5" (OuterVolumeSpecName: "kube-api-access-m75h5") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "kube-api-access-m75h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: W1125 12:27:51.233279 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod744867e9_9122_4c3d_96a1_73156d822e67.slice/crio-38c019be98aef77790cc0c2ed07ab36227b8392d16d5874aa52ae31ed42bf796 WatchSource:0}: Error finding container 38c019be98aef77790cc0c2ed07ab36227b8392d16d5874aa52ae31ed42bf796: Status 404 returned error can't find the container with id 38c019be98aef77790cc0c2ed07ab36227b8392d16d5874aa52ae31ed42bf796 Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.234883 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-scripts" (OuterVolumeSpecName: "scripts") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.235101 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.311045 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-config-data" (OuterVolumeSpecName: "config-data") pod "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" (UID: "09f1ceb0-b80a-4b11-8e9e-04cfb5377e37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329767 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329800 4715 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-dev\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329814 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m75h5\" (UniqueName: \"kubernetes.io/projected/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-kube-api-access-m75h5\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329829 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329841 4715 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329853 4715 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329864 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-logs\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329875 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329890 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.329911 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.357621 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.373096 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.431556 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:51 crc kubenswrapper[4715]: I1125 12:27:51.431609 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.038548 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"09f1ceb0-b80a-4b11-8e9e-04cfb5377e37","Type":"ContainerDied","Data":"30824e67dc73a4b4127bbe682f06e76fae36975d81cef7dfa488895b41b2d6dd"} Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.038774 4715 scope.go:117] "RemoveContainer" containerID="32c6a5c0dcde7b74f2b6d8e0dfc907a032d3fb0e7b760a6351d7387275bc12b5" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.038586 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.046153 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"744867e9-9122-4c3d-96a1-73156d822e67","Type":"ContainerStarted","Data":"07530edee3106a79181e9f377298a797bd0ab4268554e610d667e863ea5243ac"} Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.046214 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"744867e9-9122-4c3d-96a1-73156d822e67","Type":"ContainerStarted","Data":"2614dc9ec551d0a45561e58e6c7eb0fce98b7fd52f8b7bc995cefb571e801421"} Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.046225 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"744867e9-9122-4c3d-96a1-73156d822e67","Type":"ContainerStarted","Data":"0fd6ad264c902aa81ab1e0d986981efc5d432f18ed6ca2463a38cdf9a9aa8e6b"} Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.046235 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"744867e9-9122-4c3d-96a1-73156d822e67","Type":"ContainerStarted","Data":"38c019be98aef77790cc0c2ed07ab36227b8392d16d5874aa52ae31ed42bf796"} Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.127754 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=4.127719574 podStartE2EDuration="4.127719574s" podCreationTimestamp="2025-11-25 12:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:27:52.06996164 +0000 UTC m=+1122.577464661" watchObservedRunningTime="2025-11-25 12:27:52.127719574 +0000 UTC m=+1122.635222595" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.131127 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.135312 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.136406 4715 scope.go:117] "RemoveContainer" containerID="6f18b54fcfb88137d23c27bccd1816fb2d373bff598f0e9ed83a62d49b2bd66c" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.167783 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 12:27:52 crc kubenswrapper[4715]: E1125 12:27:52.168276 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-log" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.168360 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-log" Nov 25 12:27:52 crc kubenswrapper[4715]: E1125 12:27:52.168431 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-api" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.168515 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-api" Nov 25 12:27:52 crc kubenswrapper[4715]: E1125 12:27:52.168584 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-httpd" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.168645 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-httpd" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.168859 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-api" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.168944 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-log" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.169008 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" containerName="glance-httpd" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.170306 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.172934 4715 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.175308 4715 scope.go:117] "RemoveContainer" containerID="00e3e7cb4096fb97a41463ef8618ce8d229124b5e060b72d25aaf4f072ed641a" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.179022 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243058 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f191a1-ef85-4af4-895d-a021fa062fe9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243115 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bjfp\" (UniqueName: \"kubernetes.io/projected/f3f191a1-ef85-4af4-895d-a021fa062fe9-kube-api-access-7bjfp\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243361 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3f191a1-ef85-4af4-895d-a021fa062fe9-logs\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243445 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243492 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3f191a1-ef85-4af4-895d-a021fa062fe9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243530 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243621 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243716 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3f191a1-ef85-4af4-895d-a021fa062fe9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243764 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-run\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243835 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-dev\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243885 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.243931 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-sys\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345295 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3f191a1-ef85-4af4-895d-a021fa062fe9-logs\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345351 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345374 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3f191a1-ef85-4af4-895d-a021fa062fe9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345407 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345442 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3f191a1-ef85-4af4-895d-a021fa062fe9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345504 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-run\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345525 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345555 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-dev\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345590 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345640 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-run\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345647 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-sys\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-sys\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345690 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-dev\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345687 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345724 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f191a1-ef85-4af4-895d-a021fa062fe9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345736 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345756 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bjfp\" (UniqueName: \"kubernetes.io/projected/f3f191a1-ef85-4af4-895d-a021fa062fe9-kube-api-access-7bjfp\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345787 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345876 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f3f191a1-ef85-4af4-895d-a021fa062fe9-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345889 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3f191a1-ef85-4af4-895d-a021fa062fe9-logs\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345973 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.345986 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3f191a1-ef85-4af4-895d-a021fa062fe9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.351828 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3f191a1-ef85-4af4-895d-a021fa062fe9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.353599 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f191a1-ef85-4af4-895d-a021fa062fe9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.371024 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.379352 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bjfp\" (UniqueName: \"kubernetes.io/projected/f3f191a1-ef85-4af4-895d-a021fa062fe9-kube-api-access-7bjfp\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.408039 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3f191a1-ef85-4af4-895d-a021fa062fe9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.489650 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.704944 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f1ceb0-b80a-4b11-8e9e-04cfb5377e37" path="/var/lib/kubelet/pods/09f1ceb0-b80a-4b11-8e9e-04cfb5377e37/volumes" Nov 25 12:27:52 crc kubenswrapper[4715]: I1125 12:27:52.909884 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 25 12:27:53 crc kubenswrapper[4715]: I1125 12:27:53.063311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f3f191a1-ef85-4af4-895d-a021fa062fe9","Type":"ContainerStarted","Data":"013e7025479f13009c798dcc29dcb22e3d75f643e9c90a84ea6395f8ed6acf6c"} Nov 25 12:27:54 crc kubenswrapper[4715]: I1125 12:27:54.075240 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f3f191a1-ef85-4af4-895d-a021fa062fe9","Type":"ContainerStarted","Data":"e6b390643c4c7e99799243c0899423d45d6ffd8e92cacc2ed35875cf7bcde6f2"} Nov 25 12:27:54 crc kubenswrapper[4715]: I1125 12:27:54.075514 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f3f191a1-ef85-4af4-895d-a021fa062fe9","Type":"ContainerStarted","Data":"75437552a29d44b460608ac42e2947e214cb01bc67cacecc23558389e02dc227"} Nov 25 12:27:54 crc kubenswrapper[4715]: I1125 12:27:54.075524 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"f3f191a1-ef85-4af4-895d-a021fa062fe9","Type":"ContainerStarted","Data":"5389242fe06e59687833e00b9d6c52f8bd334eba2b7f74aa21330ada8df67c6e"} Nov 25 12:27:54 crc kubenswrapper[4715]: I1125 12:27:54.122639 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.122594404 podStartE2EDuration="2.122594404s" podCreationTimestamp="2025-11-25 12:27:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:27:54.109938265 +0000 UTC m=+1124.617441316" watchObservedRunningTime="2025-11-25 12:27:54.122594404 +0000 UTC m=+1124.630097425" Nov 25 12:27:58 crc kubenswrapper[4715]: I1125 12:27:58.451764 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:58 crc kubenswrapper[4715]: I1125 12:27:58.452119 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:58 crc kubenswrapper[4715]: I1125 12:27:58.452147 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:58 crc kubenswrapper[4715]: I1125 12:27:58.486959 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:58 crc kubenswrapper[4715]: I1125 12:27:58.488096 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:58 crc kubenswrapper[4715]: I1125 12:27:58.498714 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:59 crc kubenswrapper[4715]: I1125 12:27:59.155485 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:59 crc kubenswrapper[4715]: I1125 12:27:59.155553 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:59 crc kubenswrapper[4715]: I1125 12:27:59.155577 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:59 crc kubenswrapper[4715]: I1125 12:27:59.171651 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:59 crc kubenswrapper[4715]: I1125 12:27:59.174772 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:27:59 crc kubenswrapper[4715]: I1125 12:27:59.175462 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 25 12:28:02 crc kubenswrapper[4715]: I1125 12:28:02.490613 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:02 crc kubenswrapper[4715]: I1125 12:28:02.491218 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:02 crc kubenswrapper[4715]: I1125 12:28:02.491232 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:02 crc kubenswrapper[4715]: I1125 12:28:02.512840 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:02 crc kubenswrapper[4715]: I1125 12:28:02.524568 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:02 crc kubenswrapper[4715]: I1125 12:28:02.527980 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:03 crc kubenswrapper[4715]: I1125 12:28:03.196343 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:03 crc kubenswrapper[4715]: I1125 12:28:03.196407 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:03 crc kubenswrapper[4715]: I1125 12:28:03.196420 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:03 crc kubenswrapper[4715]: I1125 12:28:03.209596 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:03 crc kubenswrapper[4715]: I1125 12:28:03.209893 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:28:03 crc kubenswrapper[4715]: I1125 12:28:03.217768 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 25 12:29:17 crc kubenswrapper[4715]: I1125 12:29:17.943847 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:29:17 crc kubenswrapper[4715]: I1125 12:29:17.944590 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:29:47 crc kubenswrapper[4715]: I1125 12:29:47.944012 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:29:47 crc kubenswrapper[4715]: I1125 12:29:47.945854 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.200294 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx"] Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.206789 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.212035 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.212552 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.213838 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx"] Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.216266 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s6hl\" (UniqueName: \"kubernetes.io/projected/49b73e46-a6f8-424a-9e5e-30c345a4d376-kube-api-access-6s6hl\") pod \"collect-profiles-29401230-qdhpx\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.216345 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49b73e46-a6f8-424a-9e5e-30c345a4d376-secret-volume\") pod \"collect-profiles-29401230-qdhpx\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.216444 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49b73e46-a6f8-424a-9e5e-30c345a4d376-config-volume\") pod \"collect-profiles-29401230-qdhpx\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.245311 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj"] Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.246650 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.256037 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl"] Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.257212 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.265490 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj"] Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.275336 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl"] Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.317443 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s6hl\" (UniqueName: \"kubernetes.io/projected/49b73e46-a6f8-424a-9e5e-30c345a4d376-kube-api-access-6s6hl\") pod \"collect-profiles-29401230-qdhpx\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.317487 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49b73e46-a6f8-424a-9e5e-30c345a4d376-secret-volume\") pod \"collect-profiles-29401230-qdhpx\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.317538 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/b65a267d-d892-44eb-860d-9a0c99367cd7-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.317568 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49b73e46-a6f8-424a-9e5e-30c345a4d376-config-volume\") pod \"collect-profiles-29401230-qdhpx\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.318666 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49b73e46-a6f8-424a-9e5e-30c345a4d376-config-volume\") pod \"collect-profiles-29401230-qdhpx\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.323359 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49b73e46-a6f8-424a-9e5e-30c345a4d376-secret-volume\") pod \"collect-profiles-29401230-qdhpx\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.333269 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s6hl\" (UniqueName: \"kubernetes.io/projected/49b73e46-a6f8-424a-9e5e-30c345a4d376-kube-api-access-6s6hl\") pod \"collect-profiles-29401230-qdhpx\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.419141 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq949\" (UniqueName: \"kubernetes.io/projected/eb040786-4c96-4f37-9292-310255481d34-kube-api-access-fq949\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.419300 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eb040786-4c96-4f37-9292-310255481d34-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.419347 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.419375 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg8xp\" (UniqueName: \"kubernetes.io/projected/b65a267d-d892-44eb-860d-9a0c99367cd7-kube-api-access-xg8xp\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.419446 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.419500 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/b65a267d-d892-44eb-860d-9a0c99367cd7-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.422489 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/b65a267d-d892-44eb-860d-9a0c99367cd7-image-cache-config-data\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.438593 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.441083 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.521442 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq949\" (UniqueName: \"kubernetes.io/projected/eb040786-4c96-4f37-9292-310255481d34-kube-api-access-fq949\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.521510 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eb040786-4c96-4f37-9292-310255481d34-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.521550 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg8xp\" (UniqueName: \"kubernetes.io/projected/b65a267d-d892-44eb-860d-9a0c99367cd7-kube-api-access-xg8xp\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.529066 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eb040786-4c96-4f37-9292-310255481d34-image-cache-config-data\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.535722 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.540484 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg8xp\" (UniqueName: \"kubernetes.io/projected/b65a267d-d892-44eb-860d-9a0c99367cd7-kube-api-access-xg8xp\") pod \"glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.542066 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq949\" (UniqueName: \"kubernetes.io/projected/eb040786-4c96-4f37-9292-310255481d34-kube-api-access-fq949\") pod \"glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.566773 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.578671 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:00 crc kubenswrapper[4715]: I1125 12:30:00.966149 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx"] Nov 25 12:30:01 crc kubenswrapper[4715]: I1125 12:30:01.053433 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj"] Nov 25 12:30:01 crc kubenswrapper[4715]: I1125 12:30:01.058881 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl"] Nov 25 12:30:01 crc kubenswrapper[4715]: I1125 12:30:01.216660 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" event={"ID":"49b73e46-a6f8-424a-9e5e-30c345a4d376","Type":"ContainerStarted","Data":"6e32cfda80ec33a795480123f86d768accf5ce57ecb35394a14c9caef8ede08a"} Nov 25 12:30:02 crc kubenswrapper[4715]: I1125 12:30:02.226082 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" event={"ID":"eb040786-4c96-4f37-9292-310255481d34","Type":"ContainerStarted","Data":"eea0b4bd423c7be84bafc59eec928a371081f42116d937fef806117c373943cf"} Nov 25 12:30:02 crc kubenswrapper[4715]: I1125 12:30:02.226805 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" event={"ID":"eb040786-4c96-4f37-9292-310255481d34","Type":"ContainerStarted","Data":"d778a7febdb20a907025fd93ce3e436280fac90a4b3b51edd66f525266413884"} Nov 25 12:30:02 crc kubenswrapper[4715]: I1125 12:30:02.228521 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" event={"ID":"b65a267d-d892-44eb-860d-9a0c99367cd7","Type":"ContainerStarted","Data":"5b441ac271156389afe23bb91a215a164f803866946f8a39095f342c1c0ca5d9"} Nov 25 12:30:02 crc kubenswrapper[4715]: I1125 12:30:02.228551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" event={"ID":"b65a267d-d892-44eb-860d-9a0c99367cd7","Type":"ContainerStarted","Data":"83dbe8da88cefabddbb288d9e252e32d58339949796e9c524f322d997b184555"} Nov 25 12:30:02 crc kubenswrapper[4715]: I1125 12:30:02.230365 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" event={"ID":"49b73e46-a6f8-424a-9e5e-30c345a4d376","Type":"ContainerStarted","Data":"99a5df518224ca5d99b21824218642a269660f6e84227c8870597cc5a4ed9a53"} Nov 25 12:30:03 crc kubenswrapper[4715]: I1125 12:30:03.240974 4715 generic.go:334] "Generic (PLEG): container finished" podID="eb040786-4c96-4f37-9292-310255481d34" containerID="eea0b4bd423c7be84bafc59eec928a371081f42116d937fef806117c373943cf" exitCode=0 Nov 25 12:30:03 crc kubenswrapper[4715]: I1125 12:30:03.241028 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" event={"ID":"eb040786-4c96-4f37-9292-310255481d34","Type":"ContainerDied","Data":"eea0b4bd423c7be84bafc59eec928a371081f42116d937fef806117c373943cf"} Nov 25 12:30:03 crc kubenswrapper[4715]: I1125 12:30:03.243336 4715 generic.go:334] "Generic (PLEG): container finished" podID="49b73e46-a6f8-424a-9e5e-30c345a4d376" containerID="99a5df518224ca5d99b21824218642a269660f6e84227c8870597cc5a4ed9a53" exitCode=0 Nov 25 12:30:03 crc kubenswrapper[4715]: I1125 12:30:03.243381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" event={"ID":"49b73e46-a6f8-424a-9e5e-30c345a4d376","Type":"ContainerDied","Data":"99a5df518224ca5d99b21824218642a269660f6e84227c8870597cc5a4ed9a53"} Nov 25 12:30:03 crc kubenswrapper[4715]: I1125 12:30:03.302566 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" podStartSLOduration=3.3025492610000002 podStartE2EDuration="3.302549261s" podCreationTimestamp="2025-11-25 12:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:30:03.296443694 +0000 UTC m=+1253.803946715" watchObservedRunningTime="2025-11-25 12:30:03.302549261 +0000 UTC m=+1253.810052282" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.253855 4715 generic.go:334] "Generic (PLEG): container finished" podID="b65a267d-d892-44eb-860d-9a0c99367cd7" containerID="5b441ac271156389afe23bb91a215a164f803866946f8a39095f342c1c0ca5d9" exitCode=0 Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.253979 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" event={"ID":"b65a267d-d892-44eb-860d-9a0c99367cd7","Type":"ContainerDied","Data":"5b441ac271156389afe23bb91a215a164f803866946f8a39095f342c1c0ca5d9"} Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.548413 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.554777 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.684741 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq949\" (UniqueName: \"kubernetes.io/projected/eb040786-4c96-4f37-9292-310255481d34-kube-api-access-fq949\") pod \"eb040786-4c96-4f37-9292-310255481d34\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.684991 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eb040786-4c96-4f37-9292-310255481d34-image-cache-config-data\") pod \"eb040786-4c96-4f37-9292-310255481d34\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.685098 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49b73e46-a6f8-424a-9e5e-30c345a4d376-secret-volume\") pod \"49b73e46-a6f8-424a-9e5e-30c345a4d376\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.685245 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s6hl\" (UniqueName: \"kubernetes.io/projected/49b73e46-a6f8-424a-9e5e-30c345a4d376-kube-api-access-6s6hl\") pod \"49b73e46-a6f8-424a-9e5e-30c345a4d376\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.685336 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"eb040786-4c96-4f37-9292-310255481d34\" (UID: \"eb040786-4c96-4f37-9292-310255481d34\") " Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.685429 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49b73e46-a6f8-424a-9e5e-30c345a4d376-config-volume\") pod \"49b73e46-a6f8-424a-9e5e-30c345a4d376\" (UID: \"49b73e46-a6f8-424a-9e5e-30c345a4d376\") " Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.685831 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49b73e46-a6f8-424a-9e5e-30c345a4d376-config-volume" (OuterVolumeSpecName: "config-volume") pod "49b73e46-a6f8-424a-9e5e-30c345a4d376" (UID: "49b73e46-a6f8-424a-9e5e-30c345a4d376"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.690293 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "eb040786-4c96-4f37-9292-310255481d34" (UID: "eb040786-4c96-4f37-9292-310255481d34"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.690438 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb040786-4c96-4f37-9292-310255481d34-kube-api-access-fq949" (OuterVolumeSpecName: "kube-api-access-fq949") pod "eb040786-4c96-4f37-9292-310255481d34" (UID: "eb040786-4c96-4f37-9292-310255481d34"). InnerVolumeSpecName "kube-api-access-fq949". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.690805 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b73e46-a6f8-424a-9e5e-30c345a4d376-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "49b73e46-a6f8-424a-9e5e-30c345a4d376" (UID: "49b73e46-a6f8-424a-9e5e-30c345a4d376"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.690876 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb040786-4c96-4f37-9292-310255481d34-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "eb040786-4c96-4f37-9292-310255481d34" (UID: "eb040786-4c96-4f37-9292-310255481d34"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.692013 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b73e46-a6f8-424a-9e5e-30c345a4d376-kube-api-access-6s6hl" (OuterVolumeSpecName: "kube-api-access-6s6hl") pod "49b73e46-a6f8-424a-9e5e-30c345a4d376" (UID: "49b73e46-a6f8-424a-9e5e-30c345a4d376"). InnerVolumeSpecName "kube-api-access-6s6hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.787760 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq949\" (UniqueName: \"kubernetes.io/projected/eb040786-4c96-4f37-9292-310255481d34-kube-api-access-fq949\") on node \"crc\" DevicePath \"\"" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.787799 4715 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/eb040786-4c96-4f37-9292-310255481d34-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.787813 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49b73e46-a6f8-424a-9e5e-30c345a4d376-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.787825 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s6hl\" (UniqueName: \"kubernetes.io/projected/49b73e46-a6f8-424a-9e5e-30c345a4d376-kube-api-access-6s6hl\") on node \"crc\" DevicePath \"\"" Nov 25 12:30:04 crc kubenswrapper[4715]: I1125 12:30:04.787838 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49b73e46-a6f8-424a-9e5e-30c345a4d376-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.273920 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" event={"ID":"eb040786-4c96-4f37-9292-310255481d34","Type":"ContainerDied","Data":"d778a7febdb20a907025fd93ce3e436280fac90a4b3b51edd66f525266413884"} Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.273963 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d778a7febdb20a907025fd93ce3e436280fac90a4b3b51edd66f525266413884" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.273971 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.278589 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.278608 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401230-qdhpx" event={"ID":"49b73e46-a6f8-424a-9e5e-30c345a4d376","Type":"ContainerDied","Data":"6e32cfda80ec33a795480123f86d768accf5ce57ecb35394a14c9caef8ede08a"} Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.278659 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e32cfda80ec33a795480123f86d768accf5ce57ecb35394a14c9caef8ede08a" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.509766 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.599503 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/b65a267d-d892-44eb-860d-9a0c99367cd7-image-cache-config-data\") pod \"b65a267d-d892-44eb-860d-9a0c99367cd7\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.599551 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"b65a267d-d892-44eb-860d-9a0c99367cd7\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.599646 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg8xp\" (UniqueName: \"kubernetes.io/projected/b65a267d-d892-44eb-860d-9a0c99367cd7-kube-api-access-xg8xp\") pod \"b65a267d-d892-44eb-860d-9a0c99367cd7\" (UID: \"b65a267d-d892-44eb-860d-9a0c99367cd7\") " Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.608228 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "b65a267d-d892-44eb-860d-9a0c99367cd7" (UID: "b65a267d-d892-44eb-860d-9a0c99367cd7"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.610632 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b65a267d-d892-44eb-860d-9a0c99367cd7-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "b65a267d-d892-44eb-860d-9a0c99367cd7" (UID: "b65a267d-d892-44eb-860d-9a0c99367cd7"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.615344 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b65a267d-d892-44eb-860d-9a0c99367cd7-kube-api-access-xg8xp" (OuterVolumeSpecName: "kube-api-access-xg8xp") pod "b65a267d-d892-44eb-860d-9a0c99367cd7" (UID: "b65a267d-d892-44eb-860d-9a0c99367cd7"). InnerVolumeSpecName "kube-api-access-xg8xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.701533 4715 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/b65a267d-d892-44eb-860d-9a0c99367cd7-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:30:05 crc kubenswrapper[4715]: I1125 12:30:05.701567 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg8xp\" (UniqueName: \"kubernetes.io/projected/b65a267d-d892-44eb-860d-9a0c99367cd7-kube-api-access-xg8xp\") on node \"crc\" DevicePath \"\"" Nov 25 12:30:06 crc kubenswrapper[4715]: I1125 12:30:06.289296 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" event={"ID":"b65a267d-d892-44eb-860d-9a0c99367cd7","Type":"ContainerDied","Data":"83dbe8da88cefabddbb288d9e252e32d58339949796e9c524f322d997b184555"} Nov 25 12:30:06 crc kubenswrapper[4715]: I1125 12:30:06.290160 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83dbe8da88cefabddbb288d9e252e32d58339949796e9c524f322d997b184555" Nov 25 12:30:06 crc kubenswrapper[4715]: I1125 12:30:06.289335 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl" Nov 25 12:30:17 crc kubenswrapper[4715]: I1125 12:30:17.943303 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:30:17 crc kubenswrapper[4715]: I1125 12:30:17.943911 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:30:17 crc kubenswrapper[4715]: I1125 12:30:17.943978 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:30:17 crc kubenswrapper[4715]: I1125 12:30:17.944843 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ad3f8b219f61d21b837d289c8ec7099946d5ca2ed30ff86a4db4d5d526737939"} pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:30:17 crc kubenswrapper[4715]: I1125 12:30:17.944929 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" containerID="cri-o://ad3f8b219f61d21b837d289c8ec7099946d5ca2ed30ff86a4db4d5d526737939" gracePeriod=600 Nov 25 12:30:19 crc kubenswrapper[4715]: I1125 12:30:19.393150 4715 generic.go:334] "Generic (PLEG): container finished" podID="51676837-adc4-4424-b527-920a6528b6a2" containerID="ad3f8b219f61d21b837d289c8ec7099946d5ca2ed30ff86a4db4d5d526737939" exitCode=0 Nov 25 12:30:19 crc kubenswrapper[4715]: I1125 12:30:19.393220 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerDied","Data":"ad3f8b219f61d21b837d289c8ec7099946d5ca2ed30ff86a4db4d5d526737939"} Nov 25 12:30:19 crc kubenswrapper[4715]: I1125 12:30:19.393686 4715 scope.go:117] "RemoveContainer" containerID="244d5332aa55fb80cd40f8e16728cd8f4bf2022a5bf11e13456cf894fe4243d1" Nov 25 12:30:20 crc kubenswrapper[4715]: I1125 12:30:20.403667 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"07486165d59fb5007f467ebf84789fed491104101490863522cbea8716d593d0"} Nov 25 12:32:11 crc kubenswrapper[4715]: I1125 12:32:11.210773 4715 scope.go:117] "RemoveContainer" containerID="68706bddd9c2ca96efb4927dd5e3280d7ea7228514e773f587a18590558e1974" Nov 25 12:32:11 crc kubenswrapper[4715]: I1125 12:32:11.242411 4715 scope.go:117] "RemoveContainer" containerID="f653bc3cb656cf3aa2e877aa07917f272672cc5b2803983aade68543e112a904" Nov 25 12:32:11 crc kubenswrapper[4715]: I1125 12:32:11.300206 4715 scope.go:117] "RemoveContainer" containerID="f0b0ce2cf823279bb05351f89239a50ba7b88fb0afaf058b32afdb9472a841d7" Nov 25 12:32:47 crc kubenswrapper[4715]: I1125 12:32:47.943831 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:32:47 crc kubenswrapper[4715]: I1125 12:32:47.944456 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:33:11 crc kubenswrapper[4715]: I1125 12:33:11.373637 4715 scope.go:117] "RemoveContainer" containerID="78dac332f9b3a5cdc999e2d58c6bffe4e91b09c113a258d3cd1aa176bb7da2ea" Nov 25 12:33:17 crc kubenswrapper[4715]: I1125 12:33:17.943829 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:33:17 crc kubenswrapper[4715]: I1125 12:33:17.944559 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:33:46 crc kubenswrapper[4715]: I1125 12:33:46.082327 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-4rz8m"] Nov 25 12:33:46 crc kubenswrapper[4715]: I1125 12:33:46.088963 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-4rz8m"] Nov 25 12:33:46 crc kubenswrapper[4715]: I1125 12:33:46.096447 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl"] Nov 25 12:33:46 crc kubenswrapper[4715]: I1125 12:33:46.109400 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-dc20-account-create-update-r7gcl"] Nov 25 12:33:46 crc kubenswrapper[4715]: I1125 12:33:46.709984 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7" path="/var/lib/kubelet/pods/7e37bc21-b59a-4269-b00b-7ecf1ef2d5f7/volumes" Nov 25 12:33:46 crc kubenswrapper[4715]: I1125 12:33:46.711489 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2d64e67-9d9c-46c1-855e-05f6343e223e" path="/var/lib/kubelet/pods/d2d64e67-9d9c-46c1-855e-05f6343e223e/volumes" Nov 25 12:33:47 crc kubenswrapper[4715]: I1125 12:33:47.943604 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:33:47 crc kubenswrapper[4715]: I1125 12:33:47.943949 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:33:47 crc kubenswrapper[4715]: I1125 12:33:47.943991 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:33:47 crc kubenswrapper[4715]: I1125 12:33:47.944701 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07486165d59fb5007f467ebf84789fed491104101490863522cbea8716d593d0"} pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:33:47 crc kubenswrapper[4715]: I1125 12:33:47.944765 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" containerID="cri-o://07486165d59fb5007f467ebf84789fed491104101490863522cbea8716d593d0" gracePeriod=600 Nov 25 12:33:48 crc kubenswrapper[4715]: I1125 12:33:48.133751 4715 generic.go:334] "Generic (PLEG): container finished" podID="51676837-adc4-4424-b527-920a6528b6a2" containerID="07486165d59fb5007f467ebf84789fed491104101490863522cbea8716d593d0" exitCode=0 Nov 25 12:33:48 crc kubenswrapper[4715]: I1125 12:33:48.133804 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerDied","Data":"07486165d59fb5007f467ebf84789fed491104101490863522cbea8716d593d0"} Nov 25 12:33:48 crc kubenswrapper[4715]: I1125 12:33:48.133841 4715 scope.go:117] "RemoveContainer" containerID="ad3f8b219f61d21b837d289c8ec7099946d5ca2ed30ff86a4db4d5d526737939" Nov 25 12:33:49 crc kubenswrapper[4715]: I1125 12:33:49.147237 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerStarted","Data":"bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596"} Nov 25 12:34:02 crc kubenswrapper[4715]: I1125 12:34:02.035354 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-hspqn"] Nov 25 12:34:02 crc kubenswrapper[4715]: I1125 12:34:02.043505 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-hspqn"] Nov 25 12:34:02 crc kubenswrapper[4715]: I1125 12:34:02.704966 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2970fdeb-daab-4cdd-a29c-018c630df101" path="/var/lib/kubelet/pods/2970fdeb-daab-4cdd-a29c-018c630df101/volumes" Nov 25 12:34:09 crc kubenswrapper[4715]: I1125 12:34:09.031022 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-hw97n"] Nov 25 12:34:09 crc kubenswrapper[4715]: I1125 12:34:09.060402 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-hw97n"] Nov 25 12:34:10 crc kubenswrapper[4715]: I1125 12:34:10.704738 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77036b2e-4608-4b32-bf8c-27099d539710" path="/var/lib/kubelet/pods/77036b2e-4608-4b32-bf8c-27099d539710/volumes" Nov 25 12:34:11 crc kubenswrapper[4715]: I1125 12:34:11.450512 4715 scope.go:117] "RemoveContainer" containerID="0efe25f6b5193cd666ee0e78ff06bb1f23896d5beff658e4a459d6117df8d2ca" Nov 25 12:34:11 crc kubenswrapper[4715]: I1125 12:34:11.477848 4715 scope.go:117] "RemoveContainer" containerID="ac107488d1aed9b03ae67615decb7c97a28dd3e761da5b2c0c06b442c8609752" Nov 25 12:34:11 crc kubenswrapper[4715]: I1125 12:34:11.505436 4715 scope.go:117] "RemoveContainer" containerID="5b9c3eb7e2c24f222fd8a35655b3f27a95f3fe258f9eb0b63762658998077f01" Nov 25 12:34:11 crc kubenswrapper[4715]: I1125 12:34:11.556543 4715 scope.go:117] "RemoveContainer" containerID="2dc1c42c543a4346d8fe38356d143fb395d96a7534cbb33e570d56f74b758b39" Nov 25 12:34:11 crc kubenswrapper[4715]: I1125 12:34:11.584041 4715 scope.go:117] "RemoveContainer" containerID="a4fd561c15a66b1f930af74f7b481476cd3c3658f14a06cd278d86fc5660d247" Nov 25 12:34:11 crc kubenswrapper[4715]: I1125 12:34:11.614656 4715 scope.go:117] "RemoveContainer" containerID="bbc0ae1442d2275b48dcc3e9b65da7bdef05499cabf2547a7969573ae90d7f84" Nov 25 12:34:11 crc kubenswrapper[4715]: I1125 12:34:11.635782 4715 scope.go:117] "RemoveContainer" containerID="cca203ce6481b0cc50de66c6e121c614082f0d4fe3a52beeecd793596406a05a" Nov 25 12:34:20 crc kubenswrapper[4715]: I1125 12:34:20.927284 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8p4mv"] Nov 25 12:34:20 crc kubenswrapper[4715]: E1125 12:34:20.928421 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b65a267d-d892-44eb-860d-9a0c99367cd7" containerName="glance-cache-glance-default-external-api-0-cleaner" Nov 25 12:34:20 crc kubenswrapper[4715]: I1125 12:34:20.928437 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b65a267d-d892-44eb-860d-9a0c99367cd7" containerName="glance-cache-glance-default-external-api-0-cleaner" Nov 25 12:34:20 crc kubenswrapper[4715]: E1125 12:34:20.928462 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb040786-4c96-4f37-9292-310255481d34" containerName="glance-cache-glance-default-internal-api-0-cleaner" Nov 25 12:34:20 crc kubenswrapper[4715]: I1125 12:34:20.928468 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb040786-4c96-4f37-9292-310255481d34" containerName="glance-cache-glance-default-internal-api-0-cleaner" Nov 25 12:34:20 crc kubenswrapper[4715]: E1125 12:34:20.928497 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b73e46-a6f8-424a-9e5e-30c345a4d376" containerName="collect-profiles" Nov 25 12:34:20 crc kubenswrapper[4715]: I1125 12:34:20.928505 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b73e46-a6f8-424a-9e5e-30c345a4d376" containerName="collect-profiles" Nov 25 12:34:20 crc kubenswrapper[4715]: I1125 12:34:20.928628 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b65a267d-d892-44eb-860d-9a0c99367cd7" containerName="glance-cache-glance-default-external-api-0-cleaner" Nov 25 12:34:20 crc kubenswrapper[4715]: I1125 12:34:20.928642 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb040786-4c96-4f37-9292-310255481d34" containerName="glance-cache-glance-default-internal-api-0-cleaner" Nov 25 12:34:20 crc kubenswrapper[4715]: I1125 12:34:20.928651 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b73e46-a6f8-424a-9e5e-30c345a4d376" containerName="collect-profiles" Nov 25 12:34:20 crc kubenswrapper[4715]: I1125 12:34:20.930224 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:20 crc kubenswrapper[4715]: I1125 12:34:20.954567 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8p4mv"] Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.060279 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-catalog-content\") pod \"community-operators-8p4mv\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.060359 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-utilities\") pod \"community-operators-8p4mv\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.060399 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7qcc\" (UniqueName: \"kubernetes.io/projected/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-kube-api-access-t7qcc\") pod \"community-operators-8p4mv\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.161691 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-catalog-content\") pod \"community-operators-8p4mv\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.161782 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-utilities\") pod \"community-operators-8p4mv\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.161821 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7qcc\" (UniqueName: \"kubernetes.io/projected/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-kube-api-access-t7qcc\") pod \"community-operators-8p4mv\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.162293 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-catalog-content\") pod \"community-operators-8p4mv\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.162320 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-utilities\") pod \"community-operators-8p4mv\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.183214 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7qcc\" (UniqueName: \"kubernetes.io/projected/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-kube-api-access-t7qcc\") pod \"community-operators-8p4mv\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.267137 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:21 crc kubenswrapper[4715]: I1125 12:34:21.760951 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8p4mv"] Nov 25 12:34:22 crc kubenswrapper[4715]: I1125 12:34:22.439272 4715 generic.go:334] "Generic (PLEG): container finished" podID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerID="b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67" exitCode=0 Nov 25 12:34:22 crc kubenswrapper[4715]: I1125 12:34:22.439349 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8p4mv" event={"ID":"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78","Type":"ContainerDied","Data":"b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67"} Nov 25 12:34:22 crc kubenswrapper[4715]: I1125 12:34:22.439673 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8p4mv" event={"ID":"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78","Type":"ContainerStarted","Data":"d578c0d0d6fd0fa079905f1d0e5e2ae0bd99b4d20d5efead98022e2cbc47456f"} Nov 25 12:34:22 crc kubenswrapper[4715]: I1125 12:34:22.441907 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 12:34:23 crc kubenswrapper[4715]: I1125 12:34:23.450042 4715 generic.go:334] "Generic (PLEG): container finished" podID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerID="a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882" exitCode=0 Nov 25 12:34:23 crc kubenswrapper[4715]: I1125 12:34:23.450127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8p4mv" event={"ID":"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78","Type":"ContainerDied","Data":"a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882"} Nov 25 12:34:24 crc kubenswrapper[4715]: I1125 12:34:24.461088 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8p4mv" event={"ID":"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78","Type":"ContainerStarted","Data":"8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252"} Nov 25 12:34:24 crc kubenswrapper[4715]: I1125 12:34:24.480306 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8p4mv" podStartSLOduration=2.773446168 podStartE2EDuration="4.480289956s" podCreationTimestamp="2025-11-25 12:34:20 +0000 UTC" firstStartedPulling="2025-11-25 12:34:22.441575981 +0000 UTC m=+1512.949079022" lastFinishedPulling="2025-11-25 12:34:24.148419789 +0000 UTC m=+1514.655922810" observedRunningTime="2025-11-25 12:34:24.479463363 +0000 UTC m=+1514.986966384" watchObservedRunningTime="2025-11-25 12:34:24.480289956 +0000 UTC m=+1514.987792977" Nov 25 12:34:31 crc kubenswrapper[4715]: I1125 12:34:31.267781 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:31 crc kubenswrapper[4715]: I1125 12:34:31.269170 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:31 crc kubenswrapper[4715]: I1125 12:34:31.310754 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:31 crc kubenswrapper[4715]: I1125 12:34:31.583579 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:31 crc kubenswrapper[4715]: I1125 12:34:31.636125 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8p4mv"] Nov 25 12:34:33 crc kubenswrapper[4715]: I1125 12:34:33.543280 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8p4mv" podUID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerName="registry-server" containerID="cri-o://8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252" gracePeriod=2 Nov 25 12:34:33 crc kubenswrapper[4715]: I1125 12:34:33.947634 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.100433 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7qcc\" (UniqueName: \"kubernetes.io/projected/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-kube-api-access-t7qcc\") pod \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.100797 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-utilities\") pod \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.100846 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-catalog-content\") pod \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\" (UID: \"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78\") " Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.101590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-utilities" (OuterVolumeSpecName: "utilities") pod "9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" (UID: "9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.106068 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-kube-api-access-t7qcc" (OuterVolumeSpecName: "kube-api-access-t7qcc") pod "9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" (UID: "9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78"). InnerVolumeSpecName "kube-api-access-t7qcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.202659 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.202688 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7qcc\" (UniqueName: \"kubernetes.io/projected/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-kube-api-access-t7qcc\") on node \"crc\" DevicePath \"\"" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.553796 4715 generic.go:334] "Generic (PLEG): container finished" podID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerID="8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252" exitCode=0 Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.553840 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8p4mv" event={"ID":"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78","Type":"ContainerDied","Data":"8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252"} Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.553884 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8p4mv" event={"ID":"9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78","Type":"ContainerDied","Data":"d578c0d0d6fd0fa079905f1d0e5e2ae0bd99b4d20d5efead98022e2cbc47456f"} Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.553897 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8p4mv" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.553908 4715 scope.go:117] "RemoveContainer" containerID="8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.572344 4715 scope.go:117] "RemoveContainer" containerID="a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.589842 4715 scope.go:117] "RemoveContainer" containerID="b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.638154 4715 scope.go:117] "RemoveContainer" containerID="8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252" Nov 25 12:34:34 crc kubenswrapper[4715]: E1125 12:34:34.638709 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252\": container with ID starting with 8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252 not found: ID does not exist" containerID="8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.638812 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252"} err="failed to get container status \"8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252\": rpc error: code = NotFound desc = could not find container \"8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252\": container with ID starting with 8698da6a65ce4bd7f0fbc3cad2941f481046c01300b501c01df76f94dabdf252 not found: ID does not exist" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.638847 4715 scope.go:117] "RemoveContainer" containerID="a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882" Nov 25 12:34:34 crc kubenswrapper[4715]: E1125 12:34:34.639247 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882\": container with ID starting with a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882 not found: ID does not exist" containerID="a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.639310 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882"} err="failed to get container status \"a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882\": rpc error: code = NotFound desc = could not find container \"a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882\": container with ID starting with a11588b034ab71b0d0a4b0e3d65ee3e832114ab0aba9477624150b59f82f8882 not found: ID does not exist" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.639348 4715 scope.go:117] "RemoveContainer" containerID="b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67" Nov 25 12:34:34 crc kubenswrapper[4715]: E1125 12:34:34.639730 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67\": container with ID starting with b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67 not found: ID does not exist" containerID="b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.639808 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67"} err="failed to get container status \"b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67\": rpc error: code = NotFound desc = could not find container \"b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67\": container with ID starting with b4f60b2a437866b23e7f8ba31cad2a11a435ae5e3d8a91f3603c6b5368a36b67 not found: ID does not exist" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.751454 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" (UID: "9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.819026 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.887820 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8p4mv"] Nov 25 12:34:34 crc kubenswrapper[4715]: I1125 12:34:34.898644 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8p4mv"] Nov 25 12:34:36 crc kubenswrapper[4715]: I1125 12:34:36.702489 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" path="/var/lib/kubelet/pods/9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78/volumes" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.015796 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7jsf9"] Nov 25 12:34:41 crc kubenswrapper[4715]: E1125 12:34:41.016715 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerName="extract-content" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.016731 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerName="extract-content" Nov 25 12:34:41 crc kubenswrapper[4715]: E1125 12:34:41.016752 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerName="registry-server" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.016761 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerName="registry-server" Nov 25 12:34:41 crc kubenswrapper[4715]: E1125 12:34:41.016780 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerName="extract-utilities" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.016790 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerName="extract-utilities" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.016977 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb37233-aee9-4e4b-9b90-dd8b3ffd8f78" containerName="registry-server" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.018235 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.040348 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jsf9"] Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.117792 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-catalog-content\") pod \"redhat-operators-7jsf9\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.118013 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-utilities\") pod \"redhat-operators-7jsf9\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.118075 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgn7b\" (UniqueName: \"kubernetes.io/projected/ff2db0cf-6aba-419c-b385-ca33d70119a8-kube-api-access-pgn7b\") pod \"redhat-operators-7jsf9\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.220035 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-catalog-content\") pod \"redhat-operators-7jsf9\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.220105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-utilities\") pod \"redhat-operators-7jsf9\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.220127 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgn7b\" (UniqueName: \"kubernetes.io/projected/ff2db0cf-6aba-419c-b385-ca33d70119a8-kube-api-access-pgn7b\") pod \"redhat-operators-7jsf9\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.220946 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-catalog-content\") pod \"redhat-operators-7jsf9\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.221250 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-utilities\") pod \"redhat-operators-7jsf9\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.242072 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgn7b\" (UniqueName: \"kubernetes.io/projected/ff2db0cf-6aba-419c-b385-ca33d70119a8-kube-api-access-pgn7b\") pod \"redhat-operators-7jsf9\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.343596 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:41 crc kubenswrapper[4715]: I1125 12:34:41.775825 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jsf9"] Nov 25 12:34:42 crc kubenswrapper[4715]: I1125 12:34:42.616773 4715 generic.go:334] "Generic (PLEG): container finished" podID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerID="37f5fdcbd959112a062bb284a3966f9b4432679f2a01df8e9c058ae3dc85e9dd" exitCode=0 Nov 25 12:34:42 crc kubenswrapper[4715]: I1125 12:34:42.616857 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jsf9" event={"ID":"ff2db0cf-6aba-419c-b385-ca33d70119a8","Type":"ContainerDied","Data":"37f5fdcbd959112a062bb284a3966f9b4432679f2a01df8e9c058ae3dc85e9dd"} Nov 25 12:34:42 crc kubenswrapper[4715]: I1125 12:34:42.617329 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jsf9" event={"ID":"ff2db0cf-6aba-419c-b385-ca33d70119a8","Type":"ContainerStarted","Data":"ae3e6f14a613c8aee4e7b356974ae9d963a9565cc91c7b21540da37f2fd2a267"} Nov 25 12:34:43 crc kubenswrapper[4715]: I1125 12:34:43.626723 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jsf9" event={"ID":"ff2db0cf-6aba-419c-b385-ca33d70119a8","Type":"ContainerStarted","Data":"f6083ddb551e5fc18fac4e23cd8e9129cf914407461f397fb1f3625423dec6bd"} Nov 25 12:34:44 crc kubenswrapper[4715]: I1125 12:34:44.636032 4715 generic.go:334] "Generic (PLEG): container finished" podID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerID="f6083ddb551e5fc18fac4e23cd8e9129cf914407461f397fb1f3625423dec6bd" exitCode=0 Nov 25 12:34:44 crc kubenswrapper[4715]: I1125 12:34:44.636269 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jsf9" event={"ID":"ff2db0cf-6aba-419c-b385-ca33d70119a8","Type":"ContainerDied","Data":"f6083ddb551e5fc18fac4e23cd8e9129cf914407461f397fb1f3625423dec6bd"} Nov 25 12:34:45 crc kubenswrapper[4715]: I1125 12:34:45.644128 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jsf9" event={"ID":"ff2db0cf-6aba-419c-b385-ca33d70119a8","Type":"ContainerStarted","Data":"c61bc691c2a5f5460bb0d1449fae5f5bcca151d3a2e1c16dab766b2bf8104141"} Nov 25 12:34:45 crc kubenswrapper[4715]: I1125 12:34:45.666346 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7jsf9" podStartSLOduration=3.006772241 podStartE2EDuration="5.666304989s" podCreationTimestamp="2025-11-25 12:34:40 +0000 UTC" firstStartedPulling="2025-11-25 12:34:42.620278054 +0000 UTC m=+1533.127781115" lastFinishedPulling="2025-11-25 12:34:45.279810842 +0000 UTC m=+1535.787313863" observedRunningTime="2025-11-25 12:34:45.658795013 +0000 UTC m=+1536.166298044" watchObservedRunningTime="2025-11-25 12:34:45.666304989 +0000 UTC m=+1536.173808000" Nov 25 12:34:51 crc kubenswrapper[4715]: I1125 12:34:51.344556 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:51 crc kubenswrapper[4715]: I1125 12:34:51.346105 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:51 crc kubenswrapper[4715]: I1125 12:34:51.390706 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:51 crc kubenswrapper[4715]: I1125 12:34:51.751033 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:52 crc kubenswrapper[4715]: I1125 12:34:52.134480 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7jsf9"] Nov 25 12:34:53 crc kubenswrapper[4715]: I1125 12:34:53.720379 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7jsf9" podUID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerName="registry-server" containerID="cri-o://c61bc691c2a5f5460bb0d1449fae5f5bcca151d3a2e1c16dab766b2bf8104141" gracePeriod=2 Nov 25 12:34:54 crc kubenswrapper[4715]: I1125 12:34:54.729059 4715 generic.go:334] "Generic (PLEG): container finished" podID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerID="c61bc691c2a5f5460bb0d1449fae5f5bcca151d3a2e1c16dab766b2bf8104141" exitCode=0 Nov 25 12:34:54 crc kubenswrapper[4715]: I1125 12:34:54.729100 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jsf9" event={"ID":"ff2db0cf-6aba-419c-b385-ca33d70119a8","Type":"ContainerDied","Data":"c61bc691c2a5f5460bb0d1449fae5f5bcca151d3a2e1c16dab766b2bf8104141"} Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.284660 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.463607 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgn7b\" (UniqueName: \"kubernetes.io/projected/ff2db0cf-6aba-419c-b385-ca33d70119a8-kube-api-access-pgn7b\") pod \"ff2db0cf-6aba-419c-b385-ca33d70119a8\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.463698 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-catalog-content\") pod \"ff2db0cf-6aba-419c-b385-ca33d70119a8\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.463755 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-utilities\") pod \"ff2db0cf-6aba-419c-b385-ca33d70119a8\" (UID: \"ff2db0cf-6aba-419c-b385-ca33d70119a8\") " Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.464778 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-utilities" (OuterVolumeSpecName: "utilities") pod "ff2db0cf-6aba-419c-b385-ca33d70119a8" (UID: "ff2db0cf-6aba-419c-b385-ca33d70119a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.472608 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff2db0cf-6aba-419c-b385-ca33d70119a8-kube-api-access-pgn7b" (OuterVolumeSpecName: "kube-api-access-pgn7b") pod "ff2db0cf-6aba-419c-b385-ca33d70119a8" (UID: "ff2db0cf-6aba-419c-b385-ca33d70119a8"). InnerVolumeSpecName "kube-api-access-pgn7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.565901 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgn7b\" (UniqueName: \"kubernetes.io/projected/ff2db0cf-6aba-419c-b385-ca33d70119a8-kube-api-access-pgn7b\") on node \"crc\" DevicePath \"\"" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.565941 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.568892 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff2db0cf-6aba-419c-b385-ca33d70119a8" (UID: "ff2db0cf-6aba-419c-b385-ca33d70119a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.667020 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff2db0cf-6aba-419c-b385-ca33d70119a8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.739434 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jsf9" event={"ID":"ff2db0cf-6aba-419c-b385-ca33d70119a8","Type":"ContainerDied","Data":"ae3e6f14a613c8aee4e7b356974ae9d963a9565cc91c7b21540da37f2fd2a267"} Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.739488 4715 scope.go:117] "RemoveContainer" containerID="c61bc691c2a5f5460bb0d1449fae5f5bcca151d3a2e1c16dab766b2bf8104141" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.739490 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jsf9" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.759689 4715 scope.go:117] "RemoveContainer" containerID="f6083ddb551e5fc18fac4e23cd8e9129cf914407461f397fb1f3625423dec6bd" Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.781008 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7jsf9"] Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.786981 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7jsf9"] Nov 25 12:34:55 crc kubenswrapper[4715]: I1125 12:34:55.799584 4715 scope.go:117] "RemoveContainer" containerID="37f5fdcbd959112a062bb284a3966f9b4432679f2a01df8e9c058ae3dc85e9dd" Nov 25 12:34:56 crc kubenswrapper[4715]: I1125 12:34:56.706629 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff2db0cf-6aba-419c-b385-ca33d70119a8" path="/var/lib/kubelet/pods/ff2db0cf-6aba-419c-b385-ca33d70119a8/volumes" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.744068 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zjwwp/must-gather-fb8qj"] Nov 25 12:35:26 crc kubenswrapper[4715]: E1125 12:35:26.744944 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerName="extract-content" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.744957 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerName="extract-content" Nov 25 12:35:26 crc kubenswrapper[4715]: E1125 12:35:26.744972 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerName="extract-utilities" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.744978 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerName="extract-utilities" Nov 25 12:35:26 crc kubenswrapper[4715]: E1125 12:35:26.744995 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerName="registry-server" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.745002 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerName="registry-server" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.745221 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff2db0cf-6aba-419c-b385-ca33d70119a8" containerName="registry-server" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.746099 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zjwwp/must-gather-fb8qj"] Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.774416 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.788538 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zjwwp"/"openshift-service-ca.crt" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.788829 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zjwwp"/"kube-root-ca.crt" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.832999 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e189792e-2b07-453a-9655-e726c153fe3d-must-gather-output\") pod \"must-gather-fb8qj\" (UID: \"e189792e-2b07-453a-9655-e726c153fe3d\") " pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.833051 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wms8d\" (UniqueName: \"kubernetes.io/projected/e189792e-2b07-453a-9655-e726c153fe3d-kube-api-access-wms8d\") pod \"must-gather-fb8qj\" (UID: \"e189792e-2b07-453a-9655-e726c153fe3d\") " pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.934234 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e189792e-2b07-453a-9655-e726c153fe3d-must-gather-output\") pod \"must-gather-fb8qj\" (UID: \"e189792e-2b07-453a-9655-e726c153fe3d\") " pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.934306 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wms8d\" (UniqueName: \"kubernetes.io/projected/e189792e-2b07-453a-9655-e726c153fe3d-kube-api-access-wms8d\") pod \"must-gather-fb8qj\" (UID: \"e189792e-2b07-453a-9655-e726c153fe3d\") " pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.935044 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e189792e-2b07-453a-9655-e726c153fe3d-must-gather-output\") pod \"must-gather-fb8qj\" (UID: \"e189792e-2b07-453a-9655-e726c153fe3d\") " pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:35:26 crc kubenswrapper[4715]: I1125 12:35:26.953997 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wms8d\" (UniqueName: \"kubernetes.io/projected/e189792e-2b07-453a-9655-e726c153fe3d-kube-api-access-wms8d\") pod \"must-gather-fb8qj\" (UID: \"e189792e-2b07-453a-9655-e726c153fe3d\") " pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:35:27 crc kubenswrapper[4715]: I1125 12:35:27.106516 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:35:27 crc kubenswrapper[4715]: I1125 12:35:27.526497 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zjwwp/must-gather-fb8qj"] Nov 25 12:35:27 crc kubenswrapper[4715]: W1125 12:35:27.538410 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode189792e_2b07_453a_9655_e726c153fe3d.slice/crio-8cc7bef2905c33ad2ffa5ae363c4a7dffe282c3749f0bab53d07960cc4270343 WatchSource:0}: Error finding container 8cc7bef2905c33ad2ffa5ae363c4a7dffe282c3749f0bab53d07960cc4270343: Status 404 returned error can't find the container with id 8cc7bef2905c33ad2ffa5ae363c4a7dffe282c3749f0bab53d07960cc4270343 Nov 25 12:35:27 crc kubenswrapper[4715]: I1125 12:35:27.991605 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" event={"ID":"e189792e-2b07-453a-9655-e726c153fe3d","Type":"ContainerStarted","Data":"8cc7bef2905c33ad2ffa5ae363c4a7dffe282c3749f0bab53d07960cc4270343"} Nov 25 12:35:32 crc kubenswrapper[4715]: I1125 12:35:32.027247 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" event={"ID":"e189792e-2b07-453a-9655-e726c153fe3d","Type":"ContainerStarted","Data":"5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c"} Nov 25 12:35:32 crc kubenswrapper[4715]: I1125 12:35:32.027900 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" event={"ID":"e189792e-2b07-453a-9655-e726c153fe3d","Type":"ContainerStarted","Data":"169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599"} Nov 25 12:35:32 crc kubenswrapper[4715]: I1125 12:35:32.060730 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" podStartSLOduration=2.517353136 podStartE2EDuration="6.060706641s" podCreationTimestamp="2025-11-25 12:35:26 +0000 UTC" firstStartedPulling="2025-11-25 12:35:27.541755002 +0000 UTC m=+1578.049258023" lastFinishedPulling="2025-11-25 12:35:31.085108507 +0000 UTC m=+1581.592611528" observedRunningTime="2025-11-25 12:35:32.059539289 +0000 UTC m=+1582.567042310" watchObservedRunningTime="2025-11-25 12:35:32.060706641 +0000 UTC m=+1582.568209672" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.677382 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zw4hb"] Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.680259 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.711811 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zw4hb"] Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.833559 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8pl4\" (UniqueName: \"kubernetes.io/projected/b710e534-1c82-4d63-8609-19b188169c31-kube-api-access-q8pl4\") pod \"certified-operators-zw4hb\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.833762 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-catalog-content\") pod \"certified-operators-zw4hb\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.834035 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-utilities\") pod \"certified-operators-zw4hb\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.935324 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-utilities\") pod \"certified-operators-zw4hb\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.935396 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8pl4\" (UniqueName: \"kubernetes.io/projected/b710e534-1c82-4d63-8609-19b188169c31-kube-api-access-q8pl4\") pod \"certified-operators-zw4hb\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.935436 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-catalog-content\") pod \"certified-operators-zw4hb\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.936115 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-catalog-content\") pod \"certified-operators-zw4hb\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.936108 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-utilities\") pod \"certified-operators-zw4hb\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:03 crc kubenswrapper[4715]: I1125 12:36:03.954772 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8pl4\" (UniqueName: \"kubernetes.io/projected/b710e534-1c82-4d63-8609-19b188169c31-kube-api-access-q8pl4\") pod \"certified-operators-zw4hb\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:04 crc kubenswrapper[4715]: I1125 12:36:04.017126 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:04 crc kubenswrapper[4715]: I1125 12:36:04.529818 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zw4hb"] Nov 25 12:36:05 crc kubenswrapper[4715]: I1125 12:36:05.295952 4715 generic.go:334] "Generic (PLEG): container finished" podID="b710e534-1c82-4d63-8609-19b188169c31" containerID="86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0" exitCode=0 Nov 25 12:36:05 crc kubenswrapper[4715]: I1125 12:36:05.296097 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4hb" event={"ID":"b710e534-1c82-4d63-8609-19b188169c31","Type":"ContainerDied","Data":"86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0"} Nov 25 12:36:05 crc kubenswrapper[4715]: I1125 12:36:05.296283 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4hb" event={"ID":"b710e534-1c82-4d63-8609-19b188169c31","Type":"ContainerStarted","Data":"90b5b61e295ce61b223f31becef313d1c597520be4d0530d8203d167d16af09e"} Nov 25 12:36:06 crc kubenswrapper[4715]: I1125 12:36:06.304124 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4hb" event={"ID":"b710e534-1c82-4d63-8609-19b188169c31","Type":"ContainerStarted","Data":"b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903"} Nov 25 12:36:06 crc kubenswrapper[4715]: I1125 12:36:06.741356 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f_5d1aad28-aabe-40fb-8d6c-3d3643a6a52b/util/0.log" Nov 25 12:36:06 crc kubenswrapper[4715]: I1125 12:36:06.901828 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f_5d1aad28-aabe-40fb-8d6c-3d3643a6a52b/pull/0.log" Nov 25 12:36:06 crc kubenswrapper[4715]: I1125 12:36:06.902226 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f_5d1aad28-aabe-40fb-8d6c-3d3643a6a52b/util/0.log" Nov 25 12:36:06 crc kubenswrapper[4715]: I1125 12:36:06.947916 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f_5d1aad28-aabe-40fb-8d6c-3d3643a6a52b/pull/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.100203 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f_5d1aad28-aabe-40fb-8d6c-3d3643a6a52b/pull/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.104885 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f_5d1aad28-aabe-40fb-8d6c-3d3643a6a52b/extract/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.108150 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2f1c56fe05743b6944068527d0b6f67ed392ac583d9653ba517aab4303th76f_5d1aad28-aabe-40fb-8d6c-3d3643a6a52b/util/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.313928 4715 generic.go:334] "Generic (PLEG): container finished" podID="b710e534-1c82-4d63-8609-19b188169c31" containerID="b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903" exitCode=0 Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.313991 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4hb" event={"ID":"b710e534-1c82-4d63-8609-19b188169c31","Type":"ContainerDied","Data":"b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903"} Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.318581 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp_07d69d4e-e804-480e-86b1-1e0987585af8/util/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.409816 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp_07d69d4e-e804-480e-86b1-1e0987585af8/util/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.431426 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp_07d69d4e-e804-480e-86b1-1e0987585af8/pull/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.464470 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp_07d69d4e-e804-480e-86b1-1e0987585af8/pull/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.633238 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp_07d69d4e-e804-480e-86b1-1e0987585af8/util/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.633281 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp_07d69d4e-e804-480e-86b1-1e0987585af8/pull/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.678778 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dq2rzp_07d69d4e-e804-480e-86b1-1e0987585af8/extract/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.817603 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95_27922bfd-ecc0-49b9-b1f5-72195ceac78d/util/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.956067 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95_27922bfd-ecc0-49b9-b1f5-72195ceac78d/pull/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.982591 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95_27922bfd-ecc0-49b9-b1f5-72195ceac78d/util/0.log" Nov 25 12:36:07 crc kubenswrapper[4715]: I1125 12:36:07.997283 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95_27922bfd-ecc0-49b9-b1f5-72195ceac78d/pull/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.185775 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95_27922bfd-ecc0-49b9-b1f5-72195ceac78d/util/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.193180 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95_27922bfd-ecc0-49b9-b1f5-72195ceac78d/pull/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.214376 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_62eccd409f16609fda8c131ec699dcbc23034b86aeec0473e3f9799434twd95_27922bfd-ecc0-49b9-b1f5-72195ceac78d/extract/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.323497 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4hb" event={"ID":"b710e534-1c82-4d63-8609-19b188169c31","Type":"ContainerStarted","Data":"38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8"} Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.340695 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zw4hb" podStartSLOduration=2.906112864 podStartE2EDuration="5.340673597s" podCreationTimestamp="2025-11-25 12:36:03 +0000 UTC" firstStartedPulling="2025-11-25 12:36:05.300646377 +0000 UTC m=+1615.808149398" lastFinishedPulling="2025-11-25 12:36:07.73520711 +0000 UTC m=+1618.242710131" observedRunningTime="2025-11-25 12:36:08.338774557 +0000 UTC m=+1618.846277588" watchObservedRunningTime="2025-11-25 12:36:08.340673597 +0000 UTC m=+1618.848176618" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.434781 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm_d60e6e6e-375e-4c32-be65-21d1e1d518c6/util/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.569254 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm_d60e6e6e-375e-4c32-be65-21d1e1d518c6/util/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.644315 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm_d60e6e6e-375e-4c32-be65-21d1e1d518c6/pull/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.644438 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm_d60e6e6e-375e-4c32-be65-21d1e1d518c6/pull/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.815777 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm_d60e6e6e-375e-4c32-be65-21d1e1d518c6/util/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.840924 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm_d60e6e6e-375e-4c32-be65-21d1e1d518c6/pull/0.log" Nov 25 12:36:08 crc kubenswrapper[4715]: I1125 12:36:08.880402 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590rv8jm_d60e6e6e-375e-4c32-be65-21d1e1d518c6/extract/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.026013 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj_731e898e-a432-4696-ab69-cc1e0263558e/util/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.172128 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj_731e898e-a432-4696-ab69-cc1e0263558e/pull/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.196966 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj_731e898e-a432-4696-ab69-cc1e0263558e/util/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.197897 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj_731e898e-a432-4696-ab69-cc1e0263558e/pull/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.379949 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj_731e898e-a432-4696-ab69-cc1e0263558e/pull/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.387742 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj_731e898e-a432-4696-ab69-cc1e0263558e/util/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.393391 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_aa0b828f6f639bc0052f2eb4a1a366a08ea1b7f8aef623a055aa3d992e2cjjj_731e898e-a432-4696-ab69-cc1e0263558e/extract/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.545109 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg_68d8479f-beb5-4104-a9ca-a83b0c730f8a/util/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.679996 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg_68d8479f-beb5-4104-a9ca-a83b0c730f8a/pull/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.689549 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg_68d8479f-beb5-4104-a9ca-a83b0c730f8a/pull/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.710838 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg_68d8479f-beb5-4104-a9ca-a83b0c730f8a/util/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.824483 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg_68d8479f-beb5-4104-a9ca-a83b0c730f8a/util/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.829327 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg_68d8479f-beb5-4104-a9ca-a83b0c730f8a/extract/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.911128 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c976308faac62824ee875fa80dce4db57a79e32adb8a627dd31cdf72f6shljg_68d8479f-beb5-4104-a9ca-a83b0c730f8a/pull/0.log" Nov 25 12:36:09 crc kubenswrapper[4715]: I1125 12:36:09.913734 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs_8ff1994e-54dc-4697-9b2f-402bfb693383/util/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.087790 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs_8ff1994e-54dc-4697-9b2f-402bfb693383/util/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.088607 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs_8ff1994e-54dc-4697-9b2f-402bfb693383/pull/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.111210 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs_8ff1994e-54dc-4697-9b2f-402bfb693383/pull/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.267534 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs_8ff1994e-54dc-4697-9b2f-402bfb693383/util/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.279416 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs_8ff1994e-54dc-4697-9b2f-402bfb693383/pull/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.285992 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ef749ab7c6220a5966bbd841f607f826d6087c0d551a63f244573924aekd8rs_8ff1994e-54dc-4697-9b2f-402bfb693383/extract/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.356309 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-69cb78548c-2nfnj_5f528c96-c714-485d-8486-0b5f6efa0973/manager/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.461008 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-fkgk9_6de0e4e1-4bcd-4115-9c83-0827f880f400/registry-server/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.492834 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5cf87789bc-srhpk_116af384-b2b3-422d-84a4-38cc7affcf33/manager/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.739787 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-776cc6d755-b7wjl_2f3c70c2-5a97-479c-aed9-151ba8a3c9d8/kube-rbac-proxy/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.799747 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-776cc6d755-b7wjl_2f3c70c2-5a97-479c-aed9-151ba8a3c9d8/manager/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.802784 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-67nk8_ac6f7b65-527d-4d7c-9c36-e2144540fc4c/registry-server/0.log" Nov 25 12:36:10 crc kubenswrapper[4715]: I1125 12:36:10.924777 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-2mp82_8f728d84-3952-4b88-9fb9-565cfd25abc6/registry-server/0.log" Nov 25 12:36:11 crc kubenswrapper[4715]: I1125 12:36:11.014208 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b9d7554bc-4rbfr_9a3e4772-2d60-4498-9cdb-0a6413b4b200/manager/0.log" Nov 25 12:36:11 crc kubenswrapper[4715]: I1125 12:36:11.110948 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-4sskb_c69a3c36-24a6-4871-8d82-a811e2ed5afc/registry-server/0.log" Nov 25 12:36:11 crc kubenswrapper[4715]: I1125 12:36:11.171072 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6b6d8bbdd9-78tmp_9fb80822-9010-4c95-8c77-225fc5a92b66/manager/0.log" Nov 25 12:36:11 crc kubenswrapper[4715]: I1125 12:36:11.203397 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-wwd85_b470a68e-dbe6-4f0e-99fd-1c23ce4021bc/registry-server/0.log" Nov 25 12:36:11 crc kubenswrapper[4715]: I1125 12:36:11.308820 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-qsn2j_0508bead-e249-4ac2-84d1-741dcf4030c7/operator/0.log" Nov 25 12:36:11 crc kubenswrapper[4715]: I1125 12:36:11.368357 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-cf7p9_b1c71c09-df63-430f-80f2-74d6cdb35e1a/registry-server/0.log" Nov 25 12:36:11 crc kubenswrapper[4715]: I1125 12:36:11.521531 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7d7fdbdb55-hsz58_436fab16-f756-4564-ad51-316c4e869463/manager/0.log" Nov 25 12:36:11 crc kubenswrapper[4715]: I1125 12:36:11.562582 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-rhhbn_3ab577a0-b9fa-4213-9513-648dcb1ff3fd/registry-server/0.log" Nov 25 12:36:14 crc kubenswrapper[4715]: I1125 12:36:14.017623 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:14 crc kubenswrapper[4715]: I1125 12:36:14.018004 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:14 crc kubenswrapper[4715]: I1125 12:36:14.066937 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:14 crc kubenswrapper[4715]: I1125 12:36:14.412172 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:14 crc kubenswrapper[4715]: I1125 12:36:14.449406 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zw4hb"] Nov 25 12:36:16 crc kubenswrapper[4715]: I1125 12:36:16.375324 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zw4hb" podUID="b710e534-1c82-4d63-8609-19b188169c31" containerName="registry-server" containerID="cri-o://38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8" gracePeriod=2 Nov 25 12:36:16 crc kubenswrapper[4715]: I1125 12:36:16.780540 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:16 crc kubenswrapper[4715]: I1125 12:36:16.931375 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-utilities\") pod \"b710e534-1c82-4d63-8609-19b188169c31\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " Nov 25 12:36:16 crc kubenswrapper[4715]: I1125 12:36:16.931443 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8pl4\" (UniqueName: \"kubernetes.io/projected/b710e534-1c82-4d63-8609-19b188169c31-kube-api-access-q8pl4\") pod \"b710e534-1c82-4d63-8609-19b188169c31\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " Nov 25 12:36:16 crc kubenswrapper[4715]: I1125 12:36:16.931460 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-catalog-content\") pod \"b710e534-1c82-4d63-8609-19b188169c31\" (UID: \"b710e534-1c82-4d63-8609-19b188169c31\") " Nov 25 12:36:16 crc kubenswrapper[4715]: I1125 12:36:16.932565 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-utilities" (OuterVolumeSpecName: "utilities") pod "b710e534-1c82-4d63-8609-19b188169c31" (UID: "b710e534-1c82-4d63-8609-19b188169c31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:36:16 crc kubenswrapper[4715]: I1125 12:36:16.937126 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b710e534-1c82-4d63-8609-19b188169c31-kube-api-access-q8pl4" (OuterVolumeSpecName: "kube-api-access-q8pl4") pod "b710e534-1c82-4d63-8609-19b188169c31" (UID: "b710e534-1c82-4d63-8609-19b188169c31"). InnerVolumeSpecName "kube-api-access-q8pl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.033746 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8pl4\" (UniqueName: \"kubernetes.io/projected/b710e534-1c82-4d63-8609-19b188169c31-kube-api-access-q8pl4\") on node \"crc\" DevicePath \"\"" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.033803 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.384769 4715 generic.go:334] "Generic (PLEG): container finished" podID="b710e534-1c82-4d63-8609-19b188169c31" containerID="38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8" exitCode=0 Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.384817 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4hb" event={"ID":"b710e534-1c82-4d63-8609-19b188169c31","Type":"ContainerDied","Data":"38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8"} Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.384847 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4hb" event={"ID":"b710e534-1c82-4d63-8609-19b188169c31","Type":"ContainerDied","Data":"90b5b61e295ce61b223f31becef313d1c597520be4d0530d8203d167d16af09e"} Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.384844 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4hb" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.384864 4715 scope.go:117] "RemoveContainer" containerID="38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.403222 4715 scope.go:117] "RemoveContainer" containerID="b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.424603 4715 scope.go:117] "RemoveContainer" containerID="86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.471846 4715 scope.go:117] "RemoveContainer" containerID="38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8" Nov 25 12:36:17 crc kubenswrapper[4715]: E1125 12:36:17.473625 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8\": container with ID starting with 38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8 not found: ID does not exist" containerID="38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.473666 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8"} err="failed to get container status \"38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8\": rpc error: code = NotFound desc = could not find container \"38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8\": container with ID starting with 38be97ad1d19333f2a861df0dfcb7f8920e28b8cb6f6b356e7f14fb06d71a1a8 not found: ID does not exist" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.473695 4715 scope.go:117] "RemoveContainer" containerID="b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903" Nov 25 12:36:17 crc kubenswrapper[4715]: E1125 12:36:17.474453 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903\": container with ID starting with b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903 not found: ID does not exist" containerID="b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.474500 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903"} err="failed to get container status \"b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903\": rpc error: code = NotFound desc = could not find container \"b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903\": container with ID starting with b9b703a1b3b561eef30ddf4d5b50f23f87d4229781ce07f25f0bf04f0f58b903 not found: ID does not exist" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.474522 4715 scope.go:117] "RemoveContainer" containerID="86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0" Nov 25 12:36:17 crc kubenswrapper[4715]: E1125 12:36:17.476565 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0\": container with ID starting with 86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0 not found: ID does not exist" containerID="86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.476597 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0"} err="failed to get container status \"86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0\": rpc error: code = NotFound desc = could not find container \"86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0\": container with ID starting with 86e89bf711c3c20c1d17cf4f560993436e9d79ca8e8abae0fb57346f36e39bc0 not found: ID does not exist" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.630230 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b710e534-1c82-4d63-8609-19b188169c31" (UID: "b710e534-1c82-4d63-8609-19b188169c31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.643152 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b710e534-1c82-4d63-8609-19b188169c31-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.727059 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zw4hb"] Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.735782 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zw4hb"] Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.944110 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:36:17 crc kubenswrapper[4715]: I1125 12:36:17.944515 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:36:18 crc kubenswrapper[4715]: I1125 12:36:18.701277 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b710e534-1c82-4d63-8609-19b188169c31" path="/var/lib/kubelet/pods/b710e534-1c82-4d63-8609-19b188169c31/volumes" Nov 25 12:36:25 crc kubenswrapper[4715]: I1125 12:36:25.648614 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-8s95n_45106950-65c8-40eb-8dd2-c51a9b82765b/control-plane-machine-set-operator/0.log" Nov 25 12:36:25 crc kubenswrapper[4715]: I1125 12:36:25.798366 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4blqh_2374ca5a-3a6e-461b-9d9d-84d95a37a804/kube-rbac-proxy/0.log" Nov 25 12:36:25 crc kubenswrapper[4715]: I1125 12:36:25.824520 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4blqh_2374ca5a-3a6e-461b-9d9d-84d95a37a804/machine-api-operator/0.log" Nov 25 12:36:41 crc kubenswrapper[4715]: I1125 12:36:41.685729 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-rwp55_d193542e-7b7f-4ee2-99dc-6db50e82c1c2/kube-rbac-proxy/0.log" Nov 25 12:36:41 crc kubenswrapper[4715]: I1125 12:36:41.811066 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-rwp55_d193542e-7b7f-4ee2-99dc-6db50e82c1c2/controller/0.log" Nov 25 12:36:41 crc kubenswrapper[4715]: I1125 12:36:41.854453 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-frr-files/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.040101 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-frr-files/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.072390 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-reloader/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.112144 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-reloader/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.120418 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-metrics/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.378628 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-reloader/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.393555 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-metrics/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.433561 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-frr-files/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.452867 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-metrics/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.550906 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-reloader/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.550916 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-frr-files/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.597739 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/cp-metrics/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.645862 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/controller/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.786069 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/kube-rbac-proxy/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.814155 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/frr-metrics/0.log" Nov 25 12:36:42 crc kubenswrapper[4715]: I1125 12:36:42.896406 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/kube-rbac-proxy-frr/0.log" Nov 25 12:36:43 crc kubenswrapper[4715]: I1125 12:36:43.065902 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/reloader/0.log" Nov 25 12:36:43 crc kubenswrapper[4715]: I1125 12:36:43.136809 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ddwdz_0191af85-1aa7-4c92-a45d-fc502d5dbf8c/frr/0.log" Nov 25 12:36:43 crc kubenswrapper[4715]: I1125 12:36:43.177386 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-8msp7_874e4bd7-5cb5-47f2-b905-fe2dc002546b/frr-k8s-webhook-server/0.log" Nov 25 12:36:43 crc kubenswrapper[4715]: I1125 12:36:43.273668 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6cb7984d4b-5mkxw_691af904-3d82-40e2-946f-2c7d3fd237a2/manager/0.log" Nov 25 12:36:43 crc kubenswrapper[4715]: I1125 12:36:43.362432 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-84945578d6-hsb65_449e1cf8-4b25-4371-8e3c-505f09548785/webhook-server/0.log" Nov 25 12:36:43 crc kubenswrapper[4715]: I1125 12:36:43.428144 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-lf54k_f6028855-f32e-407c-a4d5-66e8caa6db64/kube-rbac-proxy/0.log" Nov 25 12:36:43 crc kubenswrapper[4715]: I1125 12:36:43.618389 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-lf54k_f6028855-f32e-407c-a4d5-66e8caa6db64/speaker/0.log" Nov 25 12:36:47 crc kubenswrapper[4715]: I1125 12:36:47.944014 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:36:47 crc kubenswrapper[4715]: I1125 12:36:47.944526 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:36:57 crc kubenswrapper[4715]: I1125 12:36:57.990130 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-bcbc-account-create-update-xhzzd_0aa4f353-6207-4a3c-9e0c-4d04cefdab19/mariadb-account-create-update/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.154333 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-cache-glance-default-external-api-0-cleaner-2940123ws9rl_b65a267d-d892-44eb-860d-9a0c99367cd7/glance-cache-glance-default-external-api-0-cleaner/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.211827 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-cache-glance-default-internal-api-0-cleaner-2940123gq9gj_eb040786-4c96-4f37-9292-310255481d34/glance-cache-glance-default-internal-api-0-cleaner/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.328396 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-create-4qnbn_2a847f70-4c70-4d94-8336-6c81d3697ac7/mariadb-database-create/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.367513 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-hxbr7_77bf90c9-c3bc-4691-811d-975a2717e876/glance-db-sync/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.532685 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_744867e9-9122-4c3d-96a1-73156d822e67/glance-api/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.537661 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_744867e9-9122-4c3d-96a1-73156d822e67/glance-httpd/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.585127 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_744867e9-9122-4c3d-96a1-73156d822e67/glance-log/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.693342 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_f3f191a1-ef85-4af4-895d-a021fa062fe9/glance-httpd/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.724345 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_f3f191a1-ef85-4af4-895d-a021fa062fe9/glance-api/0.log" Nov 25 12:36:58 crc kubenswrapper[4715]: I1125 12:36:58.782125 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_f3f191a1-ef85-4af4-895d-a021fa062fe9/glance-log/0.log" Nov 25 12:36:59 crc kubenswrapper[4715]: I1125 12:36:59.044467 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-7f7c6799b4-5dgxc_2b2280ed-0d22-4a77-8881-b91088f48347/keystone-api/0.log" Nov 25 12:36:59 crc kubenswrapper[4715]: I1125 12:36:59.153717 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_85183f2d-ee39-42f1-9153-9a7e6599ef2b/mysql-bootstrap/0.log" Nov 25 12:36:59 crc kubenswrapper[4715]: I1125 12:36:59.337410 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_85183f2d-ee39-42f1-9153-9a7e6599ef2b/mysql-bootstrap/0.log" Nov 25 12:36:59 crc kubenswrapper[4715]: I1125 12:36:59.390881 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_85183f2d-ee39-42f1-9153-9a7e6599ef2b/galera/0.log" Nov 25 12:36:59 crc kubenswrapper[4715]: I1125 12:36:59.623879 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_8fb56a14-d3f7-4cf3-ae0b-8aad521a3423/mysql-bootstrap/0.log" Nov 25 12:36:59 crc kubenswrapper[4715]: I1125 12:36:59.750370 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_8fb56a14-d3f7-4cf3-ae0b-8aad521a3423/galera/0.log" Nov 25 12:36:59 crc kubenswrapper[4715]: I1125 12:36:59.813783 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_8fb56a14-d3f7-4cf3-ae0b-8aad521a3423/mysql-bootstrap/0.log" Nov 25 12:36:59 crc kubenswrapper[4715]: I1125 12:36:59.965829 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_473c10ab-6523-402f-b66c-5d846a033294/mysql-bootstrap/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.187299 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_473c10ab-6523-402f-b66c-5d846a033294/galera/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.196890 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_7f48ce5a-f968-4dc7-a0bb-eff579f590af/memcached/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.208112 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_473c10ab-6523-402f-b66c-5d846a033294/mysql-bootstrap/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.334179 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_30f5ad0e-dcd3-49b0-a01f-9d26f7b8047c/openstackclient/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.393500 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_42f62681-9aba-4c32-af6b-0ec4e1503b65/setup-container/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.514128 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_42f62681-9aba-4c32-af6b-0ec4e1503b65/setup-container/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.556682 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_42f62681-9aba-4c32-af6b-0ec4e1503b65/rabbitmq/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.626173 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-jlm2n_552e79b9-1232-48cb-8f68-02488b8d34e7/proxy-httpd/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.731824 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-jlm2n_552e79b9-1232-48cb-8f68-02488b8d34e7/proxy-server/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.772363 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-ll67x_9070852d-f5ea-415a-b4f5-1c0537c628f8/swift-ring-rebalance/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.908863 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/account-auditor/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.922589 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/account-reaper/0.log" Nov 25 12:37:00 crc kubenswrapper[4715]: I1125 12:37:00.938678 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/account-replicator/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.016684 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/account-server/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.052460 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/container-auditor/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.087642 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/container-replicator/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.119524 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/container-server/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.151290 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/container-updater/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.190360 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/object-auditor/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.258882 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/object-expirer/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.280882 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/object-server/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.336199 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/object-replicator/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.344327 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/object-updater/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.415507 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/rsync/0.log" Nov 25 12:37:01 crc kubenswrapper[4715]: I1125 12:37:01.500150 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_edd3d91a-8a99-4a84-880f-75c3968b264e/swift-recon-cron/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.072230 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xjzs4_b6590884-9e6f-4390-b4c1-59902c2684ac/extract-utilities/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.244820 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xjzs4_b6590884-9e6f-4390-b4c1-59902c2684ac/extract-content/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.247530 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xjzs4_b6590884-9e6f-4390-b4c1-59902c2684ac/extract-utilities/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.276428 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xjzs4_b6590884-9e6f-4390-b4c1-59902c2684ac/extract-content/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.404232 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xjzs4_b6590884-9e6f-4390-b4c1-59902c2684ac/extract-content/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.436746 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xjzs4_b6590884-9e6f-4390-b4c1-59902c2684ac/extract-utilities/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.575761 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zp575_c4034c63-b17b-4799-a34e-4432e780c343/extract-utilities/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.713096 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xjzs4_b6590884-9e6f-4390-b4c1-59902c2684ac/registry-server/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.800163 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zp575_c4034c63-b17b-4799-a34e-4432e780c343/extract-utilities/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.829880 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zp575_c4034c63-b17b-4799-a34e-4432e780c343/extract-content/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.829962 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zp575_c4034c63-b17b-4799-a34e-4432e780c343/extract-content/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.958521 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zp575_c4034c63-b17b-4799-a34e-4432e780c343/extract-content/0.log" Nov 25 12:37:14 crc kubenswrapper[4715]: I1125 12:37:14.961487 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zp575_c4034c63-b17b-4799-a34e-4432e780c343/extract-utilities/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.148407 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455/util/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.304920 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zp575_c4034c63-b17b-4799-a34e-4432e780c343/registry-server/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.383347 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455/pull/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.388134 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455/pull/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.415127 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455/util/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.609671 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455/extract/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.614967 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455/pull/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.662140 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6lxqn5_ac0313a0-ef82-409e-8ceb-fa5ccb8b9455/util/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.765768 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-25kht_5c6da464-9fba-4591-a5b1-dfd37dde0770/marketplace-operator/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.766718 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-thpk4_44cf8567-083d-44d6-bab3-a90ee6497454/extract-utilities/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.937937 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-thpk4_44cf8567-083d-44d6-bab3-a90ee6497454/extract-utilities/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.950498 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-thpk4_44cf8567-083d-44d6-bab3-a90ee6497454/extract-content/0.log" Nov 25 12:37:15 crc kubenswrapper[4715]: I1125 12:37:15.961552 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-thpk4_44cf8567-083d-44d6-bab3-a90ee6497454/extract-content/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.102552 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-thpk4_44cf8567-083d-44d6-bab3-a90ee6497454/extract-utilities/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.120412 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-thpk4_44cf8567-083d-44d6-bab3-a90ee6497454/extract-content/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.177853 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-thpk4_44cf8567-083d-44d6-bab3-a90ee6497454/registry-server/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.313236 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w95rs_bf866d85-2d8e-4ea2-b1b8-4cd1001beb74/extract-utilities/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.404842 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w95rs_bf866d85-2d8e-4ea2-b1b8-4cd1001beb74/extract-utilities/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.427124 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w95rs_bf866d85-2d8e-4ea2-b1b8-4cd1001beb74/extract-content/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.427518 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w95rs_bf866d85-2d8e-4ea2-b1b8-4cd1001beb74/extract-content/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.610211 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w95rs_bf866d85-2d8e-4ea2-b1b8-4cd1001beb74/extract-utilities/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.615885 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w95rs_bf866d85-2d8e-4ea2-b1b8-4cd1001beb74/extract-content/0.log" Nov 25 12:37:16 crc kubenswrapper[4715]: I1125 12:37:16.897337 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w95rs_bf866d85-2d8e-4ea2-b1b8-4cd1001beb74/registry-server/0.log" Nov 25 12:37:17 crc kubenswrapper[4715]: I1125 12:37:17.943744 4715 patch_prober.go:28] interesting pod/machine-config-daemon-dk9f9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:37:17 crc kubenswrapper[4715]: I1125 12:37:17.944162 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:37:17 crc kubenswrapper[4715]: I1125 12:37:17.944273 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" Nov 25 12:37:17 crc kubenswrapper[4715]: I1125 12:37:17.945262 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596"} pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:37:17 crc kubenswrapper[4715]: I1125 12:37:17.945404 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" containerName="machine-config-daemon" containerID="cri-o://bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" gracePeriod=600 Nov 25 12:37:18 crc kubenswrapper[4715]: E1125 12:37:18.077127 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:37:18 crc kubenswrapper[4715]: I1125 12:37:18.865662 4715 generic.go:334] "Generic (PLEG): container finished" podID="51676837-adc4-4424-b527-920a6528b6a2" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" exitCode=0 Nov 25 12:37:18 crc kubenswrapper[4715]: I1125 12:37:18.865739 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" event={"ID":"51676837-adc4-4424-b527-920a6528b6a2","Type":"ContainerDied","Data":"bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596"} Nov 25 12:37:18 crc kubenswrapper[4715]: I1125 12:37:18.865805 4715 scope.go:117] "RemoveContainer" containerID="07486165d59fb5007f467ebf84789fed491104101490863522cbea8716d593d0" Nov 25 12:37:18 crc kubenswrapper[4715]: I1125 12:37:18.866578 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:37:18 crc kubenswrapper[4715]: E1125 12:37:18.866942 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:37:32 crc kubenswrapper[4715]: I1125 12:37:32.694753 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:37:32 crc kubenswrapper[4715]: E1125 12:37:32.695459 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:37:36 crc kubenswrapper[4715]: I1125 12:37:36.053462 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd"] Nov 25 12:37:36 crc kubenswrapper[4715]: I1125 12:37:36.061892 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-bcbc-account-create-update-xhzzd"] Nov 25 12:37:36 crc kubenswrapper[4715]: I1125 12:37:36.704717 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aa4f353-6207-4a3c-9e0c-4d04cefdab19" path="/var/lib/kubelet/pods/0aa4f353-6207-4a3c-9e0c-4d04cefdab19/volumes" Nov 25 12:37:37 crc kubenswrapper[4715]: I1125 12:37:37.021582 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-4qnbn"] Nov 25 12:37:37 crc kubenswrapper[4715]: I1125 12:37:37.030280 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-4qnbn"] Nov 25 12:37:38 crc kubenswrapper[4715]: I1125 12:37:38.706140 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a847f70-4c70-4d94-8336-6c81d3697ac7" path="/var/lib/kubelet/pods/2a847f70-4c70-4d94-8336-6c81d3697ac7/volumes" Nov 25 12:37:44 crc kubenswrapper[4715]: I1125 12:37:44.038535 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hxbr7"] Nov 25 12:37:44 crc kubenswrapper[4715]: I1125 12:37:44.048334 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hxbr7"] Nov 25 12:37:44 crc kubenswrapper[4715]: I1125 12:37:44.694547 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:37:44 crc kubenswrapper[4715]: E1125 12:37:44.695304 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:37:44 crc kubenswrapper[4715]: I1125 12:37:44.704888 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77bf90c9-c3bc-4691-811d-975a2717e876" path="/var/lib/kubelet/pods/77bf90c9-c3bc-4691-811d-975a2717e876/volumes" Nov 25 12:37:56 crc kubenswrapper[4715]: I1125 12:37:56.693996 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:37:56 crc kubenswrapper[4715]: E1125 12:37:56.694882 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:38:08 crc kubenswrapper[4715]: I1125 12:38:08.697895 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:38:08 crc kubenswrapper[4715]: E1125 12:38:08.699806 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:38:11 crc kubenswrapper[4715]: I1125 12:38:11.802216 4715 scope.go:117] "RemoveContainer" containerID="1f10f0aa130071966a58980829c521daba9639f4201110251d7c3e04ad0f0087" Nov 25 12:38:11 crc kubenswrapper[4715]: I1125 12:38:11.848496 4715 scope.go:117] "RemoveContainer" containerID="f76baf4c89d00d914f98431134e489645a68c5b51a98f3bd126da05052fa8407" Nov 25 12:38:11 crc kubenswrapper[4715]: I1125 12:38:11.866477 4715 scope.go:117] "RemoveContainer" containerID="fcda9cb1b3598597ada716347e9305ee89b80141884f071771dcf4703dfd173c" Nov 25 12:38:22 crc kubenswrapper[4715]: I1125 12:38:22.693385 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:38:22 crc kubenswrapper[4715]: E1125 12:38:22.694236 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:38:23 crc kubenswrapper[4715]: I1125 12:38:23.391721 4715 generic.go:334] "Generic (PLEG): container finished" podID="e189792e-2b07-453a-9655-e726c153fe3d" containerID="169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599" exitCode=0 Nov 25 12:38:23 crc kubenswrapper[4715]: I1125 12:38:23.391783 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" event={"ID":"e189792e-2b07-453a-9655-e726c153fe3d","Type":"ContainerDied","Data":"169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599"} Nov 25 12:38:23 crc kubenswrapper[4715]: I1125 12:38:23.392605 4715 scope.go:117] "RemoveContainer" containerID="169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599" Nov 25 12:38:24 crc kubenswrapper[4715]: I1125 12:38:24.018536 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zjwwp_must-gather-fb8qj_e189792e-2b07-453a-9655-e726c153fe3d/gather/0.log" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.035216 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zjwwp/must-gather-fb8qj"] Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.036164 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" podUID="e189792e-2b07-453a-9655-e726c153fe3d" containerName="copy" containerID="cri-o://5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c" gracePeriod=2 Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.048158 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zjwwp/must-gather-fb8qj"] Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.451160 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zjwwp_must-gather-fb8qj_e189792e-2b07-453a-9655-e726c153fe3d/copy/0.log" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.452295 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.453416 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zjwwp_must-gather-fb8qj_e189792e-2b07-453a-9655-e726c153fe3d/copy/0.log" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.454356 4715 generic.go:334] "Generic (PLEG): container finished" podID="e189792e-2b07-453a-9655-e726c153fe3d" containerID="5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c" exitCode=143 Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.454428 4715 scope.go:117] "RemoveContainer" containerID="5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.474422 4715 scope.go:117] "RemoveContainer" containerID="169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.520594 4715 scope.go:117] "RemoveContainer" containerID="5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c" Nov 25 12:38:31 crc kubenswrapper[4715]: E1125 12:38:31.521547 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c\": container with ID starting with 5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c not found: ID does not exist" containerID="5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.521582 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c"} err="failed to get container status \"5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c\": rpc error: code = NotFound desc = could not find container \"5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c\": container with ID starting with 5053c375a8335b981b045da92ee737c6da8c9c4376e6e0e57ffa6ae3f677b22c not found: ID does not exist" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.521603 4715 scope.go:117] "RemoveContainer" containerID="169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599" Nov 25 12:38:31 crc kubenswrapper[4715]: E1125 12:38:31.521876 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599\": container with ID starting with 169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599 not found: ID does not exist" containerID="169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.521919 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599"} err="failed to get container status \"169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599\": rpc error: code = NotFound desc = could not find container \"169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599\": container with ID starting with 169d950d24fe0a07cadb0361b1371fab9414de40a1acba1008e5135c296c4599 not found: ID does not exist" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.623890 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e189792e-2b07-453a-9655-e726c153fe3d-must-gather-output\") pod \"e189792e-2b07-453a-9655-e726c153fe3d\" (UID: \"e189792e-2b07-453a-9655-e726c153fe3d\") " Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.623982 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wms8d\" (UniqueName: \"kubernetes.io/projected/e189792e-2b07-453a-9655-e726c153fe3d-kube-api-access-wms8d\") pod \"e189792e-2b07-453a-9655-e726c153fe3d\" (UID: \"e189792e-2b07-453a-9655-e726c153fe3d\") " Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.630391 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e189792e-2b07-453a-9655-e726c153fe3d-kube-api-access-wms8d" (OuterVolumeSpecName: "kube-api-access-wms8d") pod "e189792e-2b07-453a-9655-e726c153fe3d" (UID: "e189792e-2b07-453a-9655-e726c153fe3d"). InnerVolumeSpecName "kube-api-access-wms8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.709668 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e189792e-2b07-453a-9655-e726c153fe3d-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e189792e-2b07-453a-9655-e726c153fe3d" (UID: "e189792e-2b07-453a-9655-e726c153fe3d"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.907866 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wms8d\" (UniqueName: \"kubernetes.io/projected/e189792e-2b07-453a-9655-e726c153fe3d-kube-api-access-wms8d\") on node \"crc\" DevicePath \"\"" Nov 25 12:38:31 crc kubenswrapper[4715]: I1125 12:38:31.907904 4715 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e189792e-2b07-453a-9655-e726c153fe3d-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 12:38:32 crc kubenswrapper[4715]: I1125 12:38:32.462706 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zjwwp/must-gather-fb8qj" Nov 25 12:38:32 crc kubenswrapper[4715]: I1125 12:38:32.705558 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e189792e-2b07-453a-9655-e726c153fe3d" path="/var/lib/kubelet/pods/e189792e-2b07-453a-9655-e726c153fe3d/volumes" Nov 25 12:38:36 crc kubenswrapper[4715]: I1125 12:38:36.697282 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:38:36 crc kubenswrapper[4715]: E1125 12:38:36.698275 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.901563 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qvmqz"] Nov 25 12:38:40 crc kubenswrapper[4715]: E1125 12:38:40.903230 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b710e534-1c82-4d63-8609-19b188169c31" containerName="extract-content" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.903265 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b710e534-1c82-4d63-8609-19b188169c31" containerName="extract-content" Nov 25 12:38:40 crc kubenswrapper[4715]: E1125 12:38:40.905577 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e189792e-2b07-453a-9655-e726c153fe3d" containerName="gather" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.905621 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e189792e-2b07-453a-9655-e726c153fe3d" containerName="gather" Nov 25 12:38:40 crc kubenswrapper[4715]: E1125 12:38:40.905675 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e189792e-2b07-453a-9655-e726c153fe3d" containerName="copy" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.905691 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e189792e-2b07-453a-9655-e726c153fe3d" containerName="copy" Nov 25 12:38:40 crc kubenswrapper[4715]: E1125 12:38:40.905825 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b710e534-1c82-4d63-8609-19b188169c31" containerName="registry-server" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.905847 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b710e534-1c82-4d63-8609-19b188169c31" containerName="registry-server" Nov 25 12:38:40 crc kubenswrapper[4715]: E1125 12:38:40.905894 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b710e534-1c82-4d63-8609-19b188169c31" containerName="extract-utilities" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.905912 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b710e534-1c82-4d63-8609-19b188169c31" containerName="extract-utilities" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.906284 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e189792e-2b07-453a-9655-e726c153fe3d" containerName="gather" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.906350 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b710e534-1c82-4d63-8609-19b188169c31" containerName="registry-server" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.906384 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e189792e-2b07-453a-9655-e726c153fe3d" containerName="copy" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.917569 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:40 crc kubenswrapper[4715]: I1125 12:38:40.922954 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvmqz"] Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.048243 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-catalog-content\") pod \"redhat-marketplace-qvmqz\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.048337 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-utilities\") pod \"redhat-marketplace-qvmqz\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.048393 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htvdg\" (UniqueName: \"kubernetes.io/projected/83c648b2-ac03-4b8c-b50f-1104adacbbd5-kube-api-access-htvdg\") pod \"redhat-marketplace-qvmqz\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.149544 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-catalog-content\") pod \"redhat-marketplace-qvmqz\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.149691 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-utilities\") pod \"redhat-marketplace-qvmqz\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.149784 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htvdg\" (UniqueName: \"kubernetes.io/projected/83c648b2-ac03-4b8c-b50f-1104adacbbd5-kube-api-access-htvdg\") pod \"redhat-marketplace-qvmqz\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.150153 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-catalog-content\") pod \"redhat-marketplace-qvmqz\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.150503 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-utilities\") pod \"redhat-marketplace-qvmqz\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.185259 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htvdg\" (UniqueName: \"kubernetes.io/projected/83c648b2-ac03-4b8c-b50f-1104adacbbd5-kube-api-access-htvdg\") pod \"redhat-marketplace-qvmqz\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.247596 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.467099 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvmqz"] Nov 25 12:38:41 crc kubenswrapper[4715]: I1125 12:38:41.536867 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvmqz" event={"ID":"83c648b2-ac03-4b8c-b50f-1104adacbbd5","Type":"ContainerStarted","Data":"e258d5ee0329c7441d88d84040db2a7bfeb8e5346b93093fae400fbece6fc57a"} Nov 25 12:38:42 crc kubenswrapper[4715]: I1125 12:38:42.546876 4715 generic.go:334] "Generic (PLEG): container finished" podID="83c648b2-ac03-4b8c-b50f-1104adacbbd5" containerID="5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8" exitCode=0 Nov 25 12:38:42 crc kubenswrapper[4715]: I1125 12:38:42.546928 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvmqz" event={"ID":"83c648b2-ac03-4b8c-b50f-1104adacbbd5","Type":"ContainerDied","Data":"5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8"} Nov 25 12:38:43 crc kubenswrapper[4715]: I1125 12:38:43.556088 4715 generic.go:334] "Generic (PLEG): container finished" podID="83c648b2-ac03-4b8c-b50f-1104adacbbd5" containerID="61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c" exitCode=0 Nov 25 12:38:43 crc kubenswrapper[4715]: I1125 12:38:43.556167 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvmqz" event={"ID":"83c648b2-ac03-4b8c-b50f-1104adacbbd5","Type":"ContainerDied","Data":"61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c"} Nov 25 12:38:44 crc kubenswrapper[4715]: I1125 12:38:44.567810 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvmqz" event={"ID":"83c648b2-ac03-4b8c-b50f-1104adacbbd5","Type":"ContainerStarted","Data":"290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d"} Nov 25 12:38:44 crc kubenswrapper[4715]: I1125 12:38:44.591199 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qvmqz" podStartSLOduration=3.221290946 podStartE2EDuration="4.591167277s" podCreationTimestamp="2025-11-25 12:38:40 +0000 UTC" firstStartedPulling="2025-11-25 12:38:42.550133067 +0000 UTC m=+1773.057636088" lastFinishedPulling="2025-11-25 12:38:43.920009398 +0000 UTC m=+1774.427512419" observedRunningTime="2025-11-25 12:38:44.585109703 +0000 UTC m=+1775.092612724" watchObservedRunningTime="2025-11-25 12:38:44.591167277 +0000 UTC m=+1775.098670288" Nov 25 12:38:51 crc kubenswrapper[4715]: I1125 12:38:51.248049 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:51 crc kubenswrapper[4715]: I1125 12:38:51.248752 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:51 crc kubenswrapper[4715]: I1125 12:38:51.292155 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:51 crc kubenswrapper[4715]: I1125 12:38:51.676305 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:51 crc kubenswrapper[4715]: I1125 12:38:51.696293 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:38:51 crc kubenswrapper[4715]: E1125 12:38:51.696473 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:38:53 crc kubenswrapper[4715]: I1125 12:38:53.480171 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvmqz"] Nov 25 12:38:53 crc kubenswrapper[4715]: I1125 12:38:53.629342 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qvmqz" podUID="83c648b2-ac03-4b8c-b50f-1104adacbbd5" containerName="registry-server" containerID="cri-o://290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d" gracePeriod=2 Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.095224 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.148116 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htvdg\" (UniqueName: \"kubernetes.io/projected/83c648b2-ac03-4b8c-b50f-1104adacbbd5-kube-api-access-htvdg\") pod \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.148230 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-utilities\") pod \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.148365 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-catalog-content\") pod \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\" (UID: \"83c648b2-ac03-4b8c-b50f-1104adacbbd5\") " Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.149343 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-utilities" (OuterVolumeSpecName: "utilities") pod "83c648b2-ac03-4b8c-b50f-1104adacbbd5" (UID: "83c648b2-ac03-4b8c-b50f-1104adacbbd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.155558 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83c648b2-ac03-4b8c-b50f-1104adacbbd5-kube-api-access-htvdg" (OuterVolumeSpecName: "kube-api-access-htvdg") pod "83c648b2-ac03-4b8c-b50f-1104adacbbd5" (UID: "83c648b2-ac03-4b8c-b50f-1104adacbbd5"). InnerVolumeSpecName "kube-api-access-htvdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.171881 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83c648b2-ac03-4b8c-b50f-1104adacbbd5" (UID: "83c648b2-ac03-4b8c-b50f-1104adacbbd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.250427 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htvdg\" (UniqueName: \"kubernetes.io/projected/83c648b2-ac03-4b8c-b50f-1104adacbbd5-kube-api-access-htvdg\") on node \"crc\" DevicePath \"\"" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.250463 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.250471 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83c648b2-ac03-4b8c-b50f-1104adacbbd5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.651561 4715 generic.go:334] "Generic (PLEG): container finished" podID="83c648b2-ac03-4b8c-b50f-1104adacbbd5" containerID="290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d" exitCode=0 Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.651618 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvmqz" event={"ID":"83c648b2-ac03-4b8c-b50f-1104adacbbd5","Type":"ContainerDied","Data":"290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d"} Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.651660 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvmqz" event={"ID":"83c648b2-ac03-4b8c-b50f-1104adacbbd5","Type":"ContainerDied","Data":"e258d5ee0329c7441d88d84040db2a7bfeb8e5346b93093fae400fbece6fc57a"} Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.651726 4715 scope.go:117] "RemoveContainer" containerID="290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.651951 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvmqz" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.681835 4715 scope.go:117] "RemoveContainer" containerID="61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.704813 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvmqz"] Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.708499 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvmqz"] Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.733356 4715 scope.go:117] "RemoveContainer" containerID="5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.754492 4715 scope.go:117] "RemoveContainer" containerID="290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d" Nov 25 12:38:54 crc kubenswrapper[4715]: E1125 12:38:54.754968 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d\": container with ID starting with 290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d not found: ID does not exist" containerID="290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.755006 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d"} err="failed to get container status \"290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d\": rpc error: code = NotFound desc = could not find container \"290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d\": container with ID starting with 290890955a8879bdc4ff5395ac50d6af1f2d72a58f083d6895140ba61381ea7d not found: ID does not exist" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.755032 4715 scope.go:117] "RemoveContainer" containerID="61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c" Nov 25 12:38:54 crc kubenswrapper[4715]: E1125 12:38:54.755450 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c\": container with ID starting with 61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c not found: ID does not exist" containerID="61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.755483 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c"} err="failed to get container status \"61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c\": rpc error: code = NotFound desc = could not find container \"61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c\": container with ID starting with 61a25fa091f8af2973d51d77e7f668a8949978dee1a6918075d802a76b04188c not found: ID does not exist" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.755500 4715 scope.go:117] "RemoveContainer" containerID="5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8" Nov 25 12:38:54 crc kubenswrapper[4715]: E1125 12:38:54.755756 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8\": container with ID starting with 5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8 not found: ID does not exist" containerID="5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8" Nov 25 12:38:54 crc kubenswrapper[4715]: I1125 12:38:54.755921 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8"} err="failed to get container status \"5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8\": rpc error: code = NotFound desc = could not find container \"5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8\": container with ID starting with 5b75d957c4082f45aa8d80c96604318027e30eac628f5386e9f95ba262f069b8 not found: ID does not exist" Nov 25 12:38:56 crc kubenswrapper[4715]: I1125 12:38:56.709971 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83c648b2-ac03-4b8c-b50f-1104adacbbd5" path="/var/lib/kubelet/pods/83c648b2-ac03-4b8c-b50f-1104adacbbd5/volumes" Nov 25 12:39:04 crc kubenswrapper[4715]: I1125 12:39:04.694256 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:39:04 crc kubenswrapper[4715]: E1125 12:39:04.695131 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:39:15 crc kubenswrapper[4715]: I1125 12:39:15.693751 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:39:15 crc kubenswrapper[4715]: E1125 12:39:15.694877 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:39:28 crc kubenswrapper[4715]: I1125 12:39:28.693936 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:39:28 crc kubenswrapper[4715]: E1125 12:39:28.694624 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:39:43 crc kubenswrapper[4715]: I1125 12:39:43.694760 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:39:43 crc kubenswrapper[4715]: E1125 12:39:43.696814 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:39:58 crc kubenswrapper[4715]: I1125 12:39:58.694338 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:39:58 crc kubenswrapper[4715]: E1125 12:39:58.695819 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:40:10 crc kubenswrapper[4715]: I1125 12:40:10.702136 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:40:10 crc kubenswrapper[4715]: E1125 12:40:10.702994 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:40:25 crc kubenswrapper[4715]: I1125 12:40:25.694360 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:40:25 crc kubenswrapper[4715]: E1125 12:40:25.695000 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:40:38 crc kubenswrapper[4715]: I1125 12:40:38.697959 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:40:38 crc kubenswrapper[4715]: E1125 12:40:38.698932 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:40:51 crc kubenswrapper[4715]: I1125 12:40:51.693662 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:40:51 crc kubenswrapper[4715]: E1125 12:40:51.694828 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:41:03 crc kubenswrapper[4715]: I1125 12:41:03.694169 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:41:03 crc kubenswrapper[4715]: E1125 12:41:03.695018 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:41:15 crc kubenswrapper[4715]: I1125 12:41:15.694538 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:41:15 crc kubenswrapper[4715]: E1125 12:41:15.695425 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:41:26 crc kubenswrapper[4715]: I1125 12:41:26.694511 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:41:26 crc kubenswrapper[4715]: E1125 12:41:26.695473 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:41:39 crc kubenswrapper[4715]: I1125 12:41:39.694535 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:41:39 crc kubenswrapper[4715]: E1125 12:41:39.695582 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:41:54 crc kubenswrapper[4715]: I1125 12:41:54.694767 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:41:54 crc kubenswrapper[4715]: E1125 12:41:54.695947 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" Nov 25 12:42:07 crc kubenswrapper[4715]: I1125 12:42:07.694354 4715 scope.go:117] "RemoveContainer" containerID="bc5ec6e1a644b9c48a8761564cc898b564e181d44c1a5dd39559c8df58210596" Nov 25 12:42:07 crc kubenswrapper[4715]: E1125 12:42:07.695117 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dk9f9_openshift-machine-config-operator(51676837-adc4-4424-b527-920a6528b6a2)\"" pod="openshift-machine-config-operator/machine-config-daemon-dk9f9" podUID="51676837-adc4-4424-b527-920a6528b6a2" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111322046024440 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111322047017356 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111315631016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111315631015452 5ustar corecore